var/home/core/zuul-output/0000755000175000017500000000000015072215363014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015072242633015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006453272415072242623017715 0ustar rootrootOct 10 14:44:56 crc systemd[1]: Starting Kubernetes Kubelet... Oct 10 14:44:56 crc restorecon[4753]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:56 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 10 14:44:57 crc restorecon[4753]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 10 14:44:57 crc kubenswrapper[4788]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 14:44:57 crc kubenswrapper[4788]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 10 14:44:57 crc kubenswrapper[4788]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 14:44:57 crc kubenswrapper[4788]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 14:44:57 crc kubenswrapper[4788]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 10 14:44:57 crc kubenswrapper[4788]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.967531 4788 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976644 4788 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976684 4788 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976700 4788 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976708 4788 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976715 4788 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976722 4788 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976730 4788 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976739 4788 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976746 4788 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976752 4788 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976758 4788 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976764 4788 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976769 4788 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976775 4788 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976779 4788 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976785 4788 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976790 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976796 4788 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976801 4788 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976807 4788 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976812 4788 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976818 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976824 4788 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976829 4788 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976835 4788 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976840 4788 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976847 4788 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976854 4788 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976861 4788 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976867 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976872 4788 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976878 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976884 4788 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976889 4788 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976895 4788 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976900 4788 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976905 4788 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976910 4788 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976916 4788 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976923 4788 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976928 4788 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976933 4788 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976938 4788 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976943 4788 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976948 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976953 4788 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976958 4788 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976963 4788 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976969 4788 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976975 4788 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976980 4788 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976985 4788 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976990 4788 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.976997 4788 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977003 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977017 4788 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977026 4788 feature_gate.go:330] unrecognized feature gate: Example Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977033 4788 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977040 4788 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977046 4788 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977053 4788 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977063 4788 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977070 4788 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977079 4788 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977092 4788 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977099 4788 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977106 4788 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977118 4788 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977127 4788 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977163 4788 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.977171 4788 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977321 4788 flags.go:64] FLAG: --address="0.0.0.0" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977342 4788 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977356 4788 flags.go:64] FLAG: --anonymous-auth="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977370 4788 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977380 4788 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977387 4788 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977397 4788 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977407 4788 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977459 4788 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977469 4788 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977479 4788 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977487 4788 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977494 4788 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977502 4788 flags.go:64] FLAG: --cgroup-root="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977510 4788 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977517 4788 flags.go:64] FLAG: --client-ca-file="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977525 4788 flags.go:64] FLAG: --cloud-config="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977533 4788 flags.go:64] FLAG: --cloud-provider="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977541 4788 flags.go:64] FLAG: --cluster-dns="[]" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977554 4788 flags.go:64] FLAG: --cluster-domain="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977561 4788 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977570 4788 flags.go:64] FLAG: --config-dir="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977577 4788 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977586 4788 flags.go:64] FLAG: --container-log-max-files="5" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977598 4788 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977606 4788 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977614 4788 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977622 4788 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977631 4788 flags.go:64] FLAG: --contention-profiling="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977639 4788 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977647 4788 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977659 4788 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977668 4788 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977679 4788 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977687 4788 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977695 4788 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977703 4788 flags.go:64] FLAG: --enable-load-reader="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977712 4788 flags.go:64] FLAG: --enable-server="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977720 4788 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977731 4788 flags.go:64] FLAG: --event-burst="100" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977762 4788 flags.go:64] FLAG: --event-qps="50" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977770 4788 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977778 4788 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977786 4788 flags.go:64] FLAG: --eviction-hard="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977797 4788 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977804 4788 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977812 4788 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977821 4788 flags.go:64] FLAG: --eviction-soft="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977829 4788 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977836 4788 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977843 4788 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977849 4788 flags.go:64] FLAG: --experimental-mounter-path="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977855 4788 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977862 4788 flags.go:64] FLAG: --fail-swap-on="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977868 4788 flags.go:64] FLAG: --feature-gates="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977875 4788 flags.go:64] FLAG: --file-check-frequency="20s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977881 4788 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977888 4788 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977895 4788 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977902 4788 flags.go:64] FLAG: --healthz-port="10248" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977908 4788 flags.go:64] FLAG: --help="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977914 4788 flags.go:64] FLAG: --hostname-override="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977920 4788 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977928 4788 flags.go:64] FLAG: --http-check-frequency="20s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977934 4788 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977941 4788 flags.go:64] FLAG: --image-credential-provider-config="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977947 4788 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977954 4788 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977960 4788 flags.go:64] FLAG: --image-service-endpoint="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977966 4788 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977972 4788 flags.go:64] FLAG: --kube-api-burst="100" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977979 4788 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977985 4788 flags.go:64] FLAG: --kube-api-qps="50" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977991 4788 flags.go:64] FLAG: --kube-reserved="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.977998 4788 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978004 4788 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978011 4788 flags.go:64] FLAG: --kubelet-cgroups="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978017 4788 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978023 4788 flags.go:64] FLAG: --lock-file="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978029 4788 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978035 4788 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978042 4788 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978051 4788 flags.go:64] FLAG: --log-json-split-stream="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978057 4788 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978063 4788 flags.go:64] FLAG: --log-text-split-stream="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978069 4788 flags.go:64] FLAG: --logging-format="text" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978075 4788 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978082 4788 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978088 4788 flags.go:64] FLAG: --manifest-url="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978094 4788 flags.go:64] FLAG: --manifest-url-header="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978102 4788 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978108 4788 flags.go:64] FLAG: --max-open-files="1000000" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978116 4788 flags.go:64] FLAG: --max-pods="110" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978122 4788 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978129 4788 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978159 4788 flags.go:64] FLAG: --memory-manager-policy="None" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978167 4788 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978173 4788 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978180 4788 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978186 4788 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978203 4788 flags.go:64] FLAG: --node-status-max-images="50" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978209 4788 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978216 4788 flags.go:64] FLAG: --oom-score-adj="-999" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978222 4788 flags.go:64] FLAG: --pod-cidr="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978228 4788 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978239 4788 flags.go:64] FLAG: --pod-manifest-path="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978245 4788 flags.go:64] FLAG: --pod-max-pids="-1" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978252 4788 flags.go:64] FLAG: --pods-per-core="0" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978258 4788 flags.go:64] FLAG: --port="10250" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978266 4788 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978272 4788 flags.go:64] FLAG: --provider-id="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978280 4788 flags.go:64] FLAG: --qos-reserved="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978287 4788 flags.go:64] FLAG: --read-only-port="10255" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978293 4788 flags.go:64] FLAG: --register-node="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978299 4788 flags.go:64] FLAG: --register-schedulable="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978305 4788 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978316 4788 flags.go:64] FLAG: --registry-burst="10" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978322 4788 flags.go:64] FLAG: --registry-qps="5" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978328 4788 flags.go:64] FLAG: --reserved-cpus="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978335 4788 flags.go:64] FLAG: --reserved-memory="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978343 4788 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978349 4788 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978375 4788 flags.go:64] FLAG: --rotate-certificates="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978382 4788 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978390 4788 flags.go:64] FLAG: --runonce="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978397 4788 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978403 4788 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978411 4788 flags.go:64] FLAG: --seccomp-default="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978417 4788 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978423 4788 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978431 4788 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978438 4788 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978444 4788 flags.go:64] FLAG: --storage-driver-password="root" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978450 4788 flags.go:64] FLAG: --storage-driver-secure="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978456 4788 flags.go:64] FLAG: --storage-driver-table="stats" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978463 4788 flags.go:64] FLAG: --storage-driver-user="root" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978469 4788 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978475 4788 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978482 4788 flags.go:64] FLAG: --system-cgroups="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978488 4788 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978499 4788 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978505 4788 flags.go:64] FLAG: --tls-cert-file="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978511 4788 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978519 4788 flags.go:64] FLAG: --tls-min-version="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978526 4788 flags.go:64] FLAG: --tls-private-key-file="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978532 4788 flags.go:64] FLAG: --topology-manager-policy="none" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978538 4788 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978545 4788 flags.go:64] FLAG: --topology-manager-scope="container" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978553 4788 flags.go:64] FLAG: --v="2" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978573 4788 flags.go:64] FLAG: --version="false" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978581 4788 flags.go:64] FLAG: --vmodule="" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978590 4788 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.978597 4788 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978852 4788 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978862 4788 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978868 4788 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978875 4788 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978881 4788 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978886 4788 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978891 4788 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978896 4788 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978903 4788 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978908 4788 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978913 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978919 4788 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978924 4788 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978931 4788 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978936 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978941 4788 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978946 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978951 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978956 4788 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978962 4788 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978967 4788 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978972 4788 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978977 4788 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978984 4788 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978991 4788 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.978997 4788 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979003 4788 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979009 4788 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979018 4788 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979024 4788 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979031 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979037 4788 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979042 4788 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979048 4788 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979053 4788 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979059 4788 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979065 4788 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979072 4788 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979078 4788 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979083 4788 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979094 4788 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979099 4788 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979104 4788 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979109 4788 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979114 4788 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979120 4788 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979127 4788 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979134 4788 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979162 4788 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979167 4788 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979173 4788 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979179 4788 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979184 4788 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979190 4788 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979196 4788 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979202 4788 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979207 4788 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979213 4788 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979218 4788 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979225 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979232 4788 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979238 4788 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979243 4788 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979248 4788 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979254 4788 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979259 4788 feature_gate.go:330] unrecognized feature gate: Example Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979266 4788 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979271 4788 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979277 4788 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979282 4788 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.979287 4788 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.979304 4788 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.996852 4788 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.996934 4788 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997090 4788 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997114 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997122 4788 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997132 4788 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997165 4788 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997175 4788 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997183 4788 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997192 4788 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997200 4788 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997209 4788 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997218 4788 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997226 4788 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997234 4788 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997246 4788 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997269 4788 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997280 4788 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997289 4788 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997298 4788 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997306 4788 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997316 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997324 4788 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997341 4788 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997349 4788 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997357 4788 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997365 4788 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997408 4788 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997419 4788 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997428 4788 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997436 4788 feature_gate.go:330] unrecognized feature gate: Example Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997444 4788 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997462 4788 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997470 4788 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997478 4788 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997486 4788 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997494 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997502 4788 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997513 4788 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997522 4788 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997531 4788 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997540 4788 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997549 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997557 4788 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997565 4788 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997574 4788 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997582 4788 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997615 4788 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997624 4788 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997632 4788 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997640 4788 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997648 4788 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997656 4788 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997665 4788 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997673 4788 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997681 4788 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997691 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997700 4788 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997709 4788 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997717 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997725 4788 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997733 4788 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997740 4788 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997748 4788 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997756 4788 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997764 4788 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997772 4788 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997780 4788 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997788 4788 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997796 4788 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997804 4788 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997812 4788 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.997823 4788 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: I1010 14:44:57.997842 4788 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998103 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998119 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998130 4788 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998171 4788 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998181 4788 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998193 4788 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998204 4788 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998278 4788 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998287 4788 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998295 4788 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998304 4788 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998313 4788 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998321 4788 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 10 14:44:57 crc kubenswrapper[4788]: W1010 14:44:57.998329 4788 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998337 4788 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998346 4788 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998354 4788 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998364 4788 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998372 4788 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998381 4788 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998389 4788 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998397 4788 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998405 4788 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998414 4788 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998422 4788 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998431 4788 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998438 4788 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998447 4788 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998456 4788 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998464 4788 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998472 4788 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998480 4788 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998489 4788 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998498 4788 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998506 4788 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998513 4788 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998522 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998533 4788 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998541 4788 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998549 4788 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998557 4788 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998565 4788 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998574 4788 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998581 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998590 4788 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998597 4788 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998605 4788 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998614 4788 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998622 4788 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998631 4788 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998639 4788 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998647 4788 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998656 4788 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998667 4788 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998676 4788 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998686 4788 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998695 4788 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998704 4788 feature_gate.go:330] unrecognized feature gate: Example Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998714 4788 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998723 4788 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998733 4788 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998742 4788 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998750 4788 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998758 4788 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998766 4788 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998774 4788 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998784 4788 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998793 4788 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998801 4788 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998810 4788 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:57.998820 4788 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:57.998833 4788 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.000641 4788 server.go:940] "Client rotation is on, will bootstrap in background" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.009644 4788 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.009907 4788 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.012229 4788 server.go:997] "Starting client certificate rotation" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.012292 4788 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.012561 4788 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-13 02:33:37.639649939 +0000 UTC Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.012745 4788 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1523h48m39.62691138s for next certificate rotation Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.044524 4788 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.048611 4788 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.076952 4788 log.go:25] "Validated CRI v1 runtime API" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.111379 4788 log.go:25] "Validated CRI v1 image API" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.114126 4788 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.120066 4788 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-10-14-20-35-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.120173 4788 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.144092 4788 manager.go:217] Machine: {Timestamp:2025-10-10 14:44:58.139994202 +0000 UTC m=+0.589709770 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:e666a902-fb0b-4061-82b8-ccc08d87745f BootID:f6a8a518-4f69-41df-a3e7-d3eaeece5119 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:20:5e:a8 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:20:5e:a8 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b1:65:50 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:a7:f6:df Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:80:e3:b9 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:39:3c:d5 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:e9:00:fa Speed:-1 Mtu:1496} {Name:eth10 MacAddress:c2:3c:ef:2e:db:ce Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:0e:23:d5:c7:4c:2c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.144363 4788 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.144584 4788 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.145765 4788 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.145935 4788 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.145965 4788 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.146174 4788 topology_manager.go:138] "Creating topology manager with none policy" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.146185 4788 container_manager_linux.go:303] "Creating device plugin manager" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.146663 4788 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.146691 4788 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.147350 4788 state_mem.go:36] "Initialized new in-memory state store" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.147461 4788 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.151445 4788 kubelet.go:418] "Attempting to sync node with API server" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.151469 4788 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.151485 4788 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.151498 4788 kubelet.go:324] "Adding apiserver pod source" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.151513 4788 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.155643 4788 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.156589 4788 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.159416 4788 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.160926 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.160965 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.160980 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.160994 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.161015 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.161029 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.161042 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.161077 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.161093 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.161111 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.161199 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.161219 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:58.161793 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.161927 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.129:6443: connect: connection refused" logger="UnhandledError" Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:58.162081 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.162132 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.129:6443: connect: connection refused" logger="UnhandledError" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.164959 4788 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.165487 4788 server.go:1280] "Started kubelet" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.166031 4788 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.166581 4788 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.166590 4788 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.167086 4788 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 10 14:44:58 crc systemd[1]: Started Kubernetes Kubelet. Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.169208 4788 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.169271 4788 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.169323 4788 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 01:28:57.861433156 +0000 UTC Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.169386 4788 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1258h43m59.692051698s for next certificate rotation Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.169880 4788 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.169895 4788 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.170037 4788 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.170463 4788 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.170739 4788 factory.go:55] Registering systemd factory Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.170761 4788 factory.go:221] Registration of the systemd container factory successfully Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.171120 4788 server.go:460] "Adding debug handlers to kubelet server" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.171169 4788 factory.go:153] Registering CRI-O factory Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.171206 4788 factory.go:221] Registration of the crio container factory successfully Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.171322 4788 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.171357 4788 factory.go:103] Registering Raw factory Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.171381 4788 manager.go:1196] Started watching for new ooms in manager Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.172048 4788 manager.go:319] Starting recovery of all containers Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.173723 4788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.129:6443: connect: connection refused" interval="200ms" Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:58.179254 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.179572 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.129:6443: connect: connection refused" logger="UnhandledError" Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.187661 4788 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.129:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186d2871b2a73f05 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-10 14:44:58.165452549 +0000 UTC m=+0.615168107,LastTimestamp:2025-10-10 14:44:58.165452549 +0000 UTC m=+0.615168107,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.196933 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197072 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197106 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197136 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197198 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197225 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197253 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197280 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197312 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197338 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197371 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197397 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197425 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197456 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197485 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197509 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197548 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197577 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197604 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197630 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197657 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197687 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197717 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197785 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197812 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.197840 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200361 4788 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200438 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200480 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200512 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200544 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200572 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200604 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200637 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200663 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200690 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200715 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200743 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200768 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200792 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200813 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200843 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200892 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200918 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.200963 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201006 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201036 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201061 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201089 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201115 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201188 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201212 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201289 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201326 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201353 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201428 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201459 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201488 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201523 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201550 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201576 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201602 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201625 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201652 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201685 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201711 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201738 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201766 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201793 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201818 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201844 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201874 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201902 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201931 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201955 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201977 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.201999 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202020 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202041 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202062 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202086 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202109 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202132 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202184 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202204 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202225 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202246 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202267 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202287 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202307 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202332 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202359 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202385 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202408 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202430 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202451 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202486 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202511 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202539 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202559 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202581 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202601 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202622 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202642 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202664 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202696 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202729 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202753 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202776 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202804 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202830 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202853 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202880 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202904 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202931 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202954 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.202983 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203005 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203025 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203085 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203106 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203129 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203181 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203204 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203230 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203250 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203271 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203293 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203321 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203343 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203364 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203391 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203462 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203495 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203515 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203537 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203563 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203587 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203614 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203639 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203664 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203686 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203706 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203726 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203748 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203769 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203789 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203809 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203829 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203848 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203869 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203889 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203909 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203928 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203947 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203967 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.203987 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204008 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204028 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204047 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204069 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204118 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204167 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204188 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204212 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204233 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204273 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204292 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204312 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204333 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204352 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204373 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204483 4788 manager.go:324] Recovery completed Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204899 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204925 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204943 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204965 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.204985 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205004 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205024 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205042 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205062 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205080 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205098 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205118 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205137 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205184 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205203 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205220 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205237 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205258 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205282 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205302 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205323 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205343 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205360 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205378 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205397 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205414 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205432 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205458 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205478 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205496 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205514 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205534 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205551 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205569 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205587 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205604 4788 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205623 4788 reconstruct.go:97] "Volume reconstruction finished" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.205636 4788 reconciler.go:26] "Reconciler: start to sync state" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.213510 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.216320 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.216369 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.216379 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.218378 4788 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.218458 4788 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.218597 4788 state_mem.go:36] "Initialized new in-memory state store" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.229886 4788 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.232474 4788 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.232543 4788 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.232586 4788 kubelet.go:2335] "Starting kubelet main sync loop" Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.232779 4788 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:58.233252 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.233309 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.129:6443: connect: connection refused" logger="UnhandledError" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.240325 4788 policy_none.go:49] "None policy: Start" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.241387 4788 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.241412 4788 state_mem.go:35] "Initializing new in-memory state store" Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.270607 4788 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.300955 4788 manager.go:334] "Starting Device Plugin manager" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.301002 4788 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.301017 4788 server.go:79] "Starting device plugin registration server" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.301557 4788 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.301571 4788 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.302265 4788 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.302357 4788 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.302368 4788 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.310537 4788 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.333738 4788 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.333948 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.335413 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.335468 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.335482 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.335708 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.336385 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.336487 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337226 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337254 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337267 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337391 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337532 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337586 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337791 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337865 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.337888 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338116 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338163 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338177 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338343 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338383 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338415 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338427 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338463 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.338520 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339119 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339166 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339181 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339311 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339459 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339506 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339539 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339560 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.339575 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.340765 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.340789 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.340803 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.340829 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.340861 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.340878 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.342040 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.342096 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.343245 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.343288 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.343306 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.374643 4788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.129:6443: connect: connection refused" interval="400ms" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.402026 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.404259 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.404308 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.404329 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.404366 4788 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.404968 4788 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.129:6443: connect: connection refused" node="crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408324 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408383 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408424 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408457 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408549 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408600 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408639 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408672 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408755 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.408886 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.409281 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.409342 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.409382 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.409422 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.409456 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.510570 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511025 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511051 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511076 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511099 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511180 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511102 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.510809 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511246 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511256 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511321 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511333 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511407 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511496 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511555 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511578 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511605 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511601 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511662 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511695 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511503 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511664 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511763 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511797 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511829 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511846 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511863 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511898 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511912 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.511990 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.605129 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.608578 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.608647 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.608665 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.608702 4788 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.609400 4788 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.129:6443: connect: connection refused" node="crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.670614 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.692250 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.702351 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.722630 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: I1010 14:44:58.730883 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:58.736231 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-68690c874afef5f368e1bae37acdce661d2118c1ca0a22823e36b2429e4cea3c WatchSource:0}: Error finding container 68690c874afef5f368e1bae37acdce661d2118c1ca0a22823e36b2429e4cea3c: Status 404 returned error can't find the container with id 68690c874afef5f368e1bae37acdce661d2118c1ca0a22823e36b2429e4cea3c Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:58.739233 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c272739415925b0eb4b4f3e578cd8f1b15b0f2b1ed0ca1a3113a434f51e3c44a WatchSource:0}: Error finding container c272739415925b0eb4b4f3e578cd8f1b15b0f2b1ed0ca1a3113a434f51e3c44a: Status 404 returned error can't find the container with id c272739415925b0eb4b4f3e578cd8f1b15b0f2b1ed0ca1a3113a434f51e3c44a Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:58.757503 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-518c92e3a37cd076c6267479e17fd5403d96a8d1935e4be59ff6131d03eafc7d WatchSource:0}: Error finding container 518c92e3a37cd076c6267479e17fd5403d96a8d1935e4be59ff6131d03eafc7d: Status 404 returned error can't find the container with id 518c92e3a37cd076c6267479e17fd5403d96a8d1935e4be59ff6131d03eafc7d Oct 10 14:44:58 crc kubenswrapper[4788]: W1010 14:44:58.759415 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-246623c90d580cbdb739eeddf6cb2dab03a02ddbfead702e0507c0b23770a3f5 WatchSource:0}: Error finding container 246623c90d580cbdb739eeddf6cb2dab03a02ddbfead702e0507c0b23770a3f5: Status 404 returned error can't find the container with id 246623c90d580cbdb739eeddf6cb2dab03a02ddbfead702e0507c0b23770a3f5 Oct 10 14:44:58 crc kubenswrapper[4788]: E1010 14:44:58.776103 4788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.129:6443: connect: connection refused" interval="800ms" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.009550 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.011271 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.011337 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.011352 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.011389 4788 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 14:44:59 crc kubenswrapper[4788]: E1010 14:44:59.012132 4788 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.129:6443: connect: connection refused" node="crc" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.167738 4788 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.237652 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3123f2c0a16f7e6b19f532f5a81336bf231c28ea06d0ae5870714c7ddb85b265"} Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.239945 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c272739415925b0eb4b4f3e578cd8f1b15b0f2b1ed0ca1a3113a434f51e3c44a"} Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.241111 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"68690c874afef5f368e1bae37acdce661d2118c1ca0a22823e36b2429e4cea3c"} Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.242033 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"246623c90d580cbdb739eeddf6cb2dab03a02ddbfead702e0507c0b23770a3f5"} Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.243043 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"518c92e3a37cd076c6267479e17fd5403d96a8d1935e4be59ff6131d03eafc7d"} Oct 10 14:44:59 crc kubenswrapper[4788]: W1010 14:44:59.559520 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:59 crc kubenswrapper[4788]: E1010 14:44:59.559630 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.129:6443: connect: connection refused" logger="UnhandledError" Oct 10 14:44:59 crc kubenswrapper[4788]: E1010 14:44:59.577830 4788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.129:6443: connect: connection refused" interval="1.6s" Oct 10 14:44:59 crc kubenswrapper[4788]: W1010 14:44:59.671129 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:59 crc kubenswrapper[4788]: E1010 14:44:59.671410 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.129:6443: connect: connection refused" logger="UnhandledError" Oct 10 14:44:59 crc kubenswrapper[4788]: W1010 14:44:59.716822 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:59 crc kubenswrapper[4788]: E1010 14:44:59.717004 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.129:6443: connect: connection refused" logger="UnhandledError" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.814001 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.816085 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.816244 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.816275 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:44:59 crc kubenswrapper[4788]: I1010 14:44:59.816331 4788 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 14:44:59 crc kubenswrapper[4788]: E1010 14:44:59.817289 4788 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.129:6443: connect: connection refused" node="crc" Oct 10 14:44:59 crc kubenswrapper[4788]: W1010 14:44:59.821175 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:44:59 crc kubenswrapper[4788]: E1010 14:44:59.821262 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.129:6443: connect: connection refused" logger="UnhandledError" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.167270 4788 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.248121 4788 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="50c63c096a47b8a7a69372a1d6da59145d6357a732c4a10c6ba0360e58ee9f89" exitCode=0 Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.248275 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.248256 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"50c63c096a47b8a7a69372a1d6da59145d6357a732c4a10c6ba0360e58ee9f89"} Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.250074 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.250126 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.250172 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.251833 4788 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1" exitCode=0 Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.252080 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.252819 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1"} Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.253489 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.253529 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.253546 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.257043 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1"} Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.257085 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03"} Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.257102 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082"} Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.257117 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a"} Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.257121 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.258128 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.258186 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.258202 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.265914 4788 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc" exitCode=0 Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.265999 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc"} Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.266069 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.267331 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.267372 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.267384 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.275600 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.276935 4788 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d" exitCode=0 Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.276992 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d"} Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.277113 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.278965 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.278986 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.278996 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.279621 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.279650 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.279660 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:00 crc kubenswrapper[4788]: I1010 14:45:00.542183 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.167503 4788 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.129:6443: connect: connection refused Oct 10 14:45:01 crc kubenswrapper[4788]: E1010 14:45:01.179338 4788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.129:6443: connect: connection refused" interval="3.2s" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.284510 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"421443fc8d116be5e714237a5ab78e78cedbb9294dde102f4dcf88d69af814fc"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.284654 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.286323 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.286358 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.286369 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.294324 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.294370 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.294382 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.294473 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.295247 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.295274 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.295287 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.298980 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.299042 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.299061 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.299074 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.300757 4788 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e" exitCode=0 Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.300972 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.301344 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e"} Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.301321 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.302427 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.302458 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.302472 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.302578 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.302592 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.302604 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.418027 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.420481 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.420525 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.420536 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:01 crc kubenswrapper[4788]: I1010 14:45:01.420566 4788 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 14:45:01 crc kubenswrapper[4788]: E1010 14:45:01.421232 4788 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.129:6443: connect: connection refused" node="crc" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.017052 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.313902 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63"} Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.314224 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.316806 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.316905 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.316928 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.320614 4788 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8" exitCode=0 Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.320727 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8"} Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.320831 4788 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.320901 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.320901 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.320917 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.320925 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323443 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323474 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323527 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323550 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323569 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323593 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323610 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323702 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323720 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323736 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323492 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.323802 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.637166 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:45:02 crc kubenswrapper[4788]: I1010 14:45:02.648994 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.328015 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.328267 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974"} Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.328327 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654"} Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.328337 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd"} Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.328347 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859"} Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.328387 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.328496 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.329780 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.329830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.329848 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.329932 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.329964 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.329974 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.375758 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:45:03 crc kubenswrapper[4788]: I1010 14:45:03.952907 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.222681 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.333921 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c"} Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.333981 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.333981 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.334070 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335055 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335063 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335082 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335093 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335101 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335116 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335390 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335419 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.335429 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.622314 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.623451 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.623490 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.623505 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:04 crc kubenswrapper[4788]: I1010 14:45:04.623529 4788 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.017467 4788 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.017583 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.335764 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.335882 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.335889 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340198 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340244 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340258 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340279 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340302 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340309 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340304 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340459 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.340485 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:05 crc kubenswrapper[4788]: I1010 14:45:05.542990 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 10 14:45:06 crc kubenswrapper[4788]: I1010 14:45:06.338866 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:06 crc kubenswrapper[4788]: I1010 14:45:06.339973 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:06 crc kubenswrapper[4788]: I1010 14:45:06.340024 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:06 crc kubenswrapper[4788]: I1010 14:45:06.340036 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:08 crc kubenswrapper[4788]: E1010 14:45:08.310643 4788 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 10 14:45:08 crc kubenswrapper[4788]: I1010 14:45:08.514288 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:45:08 crc kubenswrapper[4788]: I1010 14:45:08.514569 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:08 crc kubenswrapper[4788]: I1010 14:45:08.516113 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:08 crc kubenswrapper[4788]: I1010 14:45:08.516247 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:08 crc kubenswrapper[4788]: I1010 14:45:08.516267 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:10 crc kubenswrapper[4788]: I1010 14:45:10.518618 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 10 14:45:10 crc kubenswrapper[4788]: I1010 14:45:10.518806 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:10 crc kubenswrapper[4788]: I1010 14:45:10.519773 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:10 crc kubenswrapper[4788]: I1010 14:45:10.519802 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:10 crc kubenswrapper[4788]: I1010 14:45:10.519814 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:11 crc kubenswrapper[4788]: I1010 14:45:11.463832 4788 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 10 14:45:11 crc kubenswrapper[4788]: I1010 14:45:11.463970 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 10 14:45:12 crc kubenswrapper[4788]: I1010 14:45:12.168032 4788 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 10 14:45:12 crc kubenswrapper[4788]: W1010 14:45:12.179341 4788 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 10 14:45:12 crc kubenswrapper[4788]: I1010 14:45:12.179426 4788 trace.go:236] Trace[1106864723]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 14:45:02.177) (total time: 10001ms): Oct 10 14:45:12 crc kubenswrapper[4788]: Trace[1106864723]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:45:12.179) Oct 10 14:45:12 crc kubenswrapper[4788]: Trace[1106864723]: [10.00163142s] [10.00163142s] END Oct 10 14:45:12 crc kubenswrapper[4788]: E1010 14:45:12.179458 4788 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 10 14:45:12 crc kubenswrapper[4788]: I1010 14:45:12.326875 4788 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 10 14:45:12 crc kubenswrapper[4788]: I1010 14:45:12.327199 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 10 14:45:12 crc kubenswrapper[4788]: I1010 14:45:12.337455 4788 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 10 14:45:12 crc kubenswrapper[4788]: I1010 14:45:12.337552 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 10 14:45:13 crc kubenswrapper[4788]: I1010 14:45:13.386335 4788 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]log ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]etcd ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/generic-apiserver-start-informers ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/priority-and-fairness-filter ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-apiextensions-informers ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-apiextensions-controllers ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/crd-informer-synced ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-system-namespaces-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 10 14:45:13 crc kubenswrapper[4788]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/bootstrap-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/start-kube-aggregator-informers ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/apiservice-registration-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/apiservice-discovery-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]autoregister-completion ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/apiservice-openapi-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 10 14:45:13 crc kubenswrapper[4788]: livez check failed Oct 10 14:45:13 crc kubenswrapper[4788]: I1010 14:45:13.386437 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:45:14 crc kubenswrapper[4788]: I1010 14:45:14.230334 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:45:14 crc kubenswrapper[4788]: I1010 14:45:14.230542 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:14 crc kubenswrapper[4788]: I1010 14:45:14.232022 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:14 crc kubenswrapper[4788]: I1010 14:45:14.232066 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:14 crc kubenswrapper[4788]: I1010 14:45:14.232078 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:15 crc kubenswrapper[4788]: I1010 14:45:15.018920 4788 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 14:45:15 crc kubenswrapper[4788]: I1010 14:45:15.019001 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 14:45:15 crc kubenswrapper[4788]: I1010 14:45:15.544174 4788 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 10 14:45:17 crc kubenswrapper[4788]: E1010 14:45:17.316544 4788 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 10 14:45:17 crc kubenswrapper[4788]: I1010 14:45:17.318494 4788 trace.go:236] Trace[1865121134]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 14:45:02.621) (total time: 14696ms): Oct 10 14:45:17 crc kubenswrapper[4788]: Trace[1865121134]: ---"Objects listed" error: 14696ms (14:45:17.318) Oct 10 14:45:17 crc kubenswrapper[4788]: Trace[1865121134]: [14.696768691s] [14.696768691s] END Oct 10 14:45:17 crc kubenswrapper[4788]: I1010 14:45:17.318535 4788 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 10 14:45:17 crc kubenswrapper[4788]: I1010 14:45:17.318607 4788 trace.go:236] Trace[935170408]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 14:45:02.787) (total time: 14531ms): Oct 10 14:45:17 crc kubenswrapper[4788]: Trace[935170408]: ---"Objects listed" error: 14531ms (14:45:17.318) Oct 10 14:45:17 crc kubenswrapper[4788]: Trace[935170408]: [14.531348176s] [14.531348176s] END Oct 10 14:45:17 crc kubenswrapper[4788]: I1010 14:45:17.318637 4788 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 10 14:45:17 crc kubenswrapper[4788]: I1010 14:45:17.319093 4788 trace.go:236] Trace[702509315]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Oct-2025 14:45:02.833) (total time: 14485ms): Oct 10 14:45:17 crc kubenswrapper[4788]: Trace[702509315]: ---"Objects listed" error: 14485ms (14:45:17.319) Oct 10 14:45:17 crc kubenswrapper[4788]: Trace[702509315]: [14.485878092s] [14.485878092s] END Oct 10 14:45:17 crc kubenswrapper[4788]: I1010 14:45:17.319121 4788 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 10 14:45:17 crc kubenswrapper[4788]: I1010 14:45:17.320520 4788 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 10 14:45:17 crc kubenswrapper[4788]: E1010 14:45:17.324739 4788 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.160766 4788 apiserver.go:52] "Watching apiserver" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.163506 4788 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.163896 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-szmv7","openshift-multus/multus-rzgfk","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-cwgns","openshift-dns/node-resolver-fdhpt","openshift-machine-config-operator/machine-config-daemon-dxfdf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.164332 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.164373 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.164481 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.164736 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.164783 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.165387 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.165434 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.165496 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.166252 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.166321 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.166639 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.167070 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fdhpt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.167477 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.167523 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.169585 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.169642 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.169821 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.169828 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.170005 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.170041 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.170492 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.170673 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.170726 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.170860 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.171297 4788 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.171760 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.171879 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.171898 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.171955 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.172112 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.172133 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.172229 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.172289 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.172390 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.173915 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174131 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174318 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174335 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174424 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174549 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174662 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174726 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174816 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174667 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174891 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.174913 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.187198 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.203925 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.214169 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.229830 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.229927 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.229992 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230045 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230097 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230179 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230230 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230288 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230344 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230421 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230480 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230543 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230604 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230661 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230725 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230785 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230885 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230964 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231035 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231076 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231112 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231189 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231232 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231267 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231310 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231347 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231383 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231436 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231488 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231523 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231565 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231617 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231694 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231767 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231826 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231881 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231934 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232001 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232059 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232111 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232204 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232263 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230579 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232312 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230672 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230880 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.230914 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231198 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231316 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231310 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231310 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231348 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232467 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231641 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.231723 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232163 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232178 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232815 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232890 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232986 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233152 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233512 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233534 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233439 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233621 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.232368 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233892 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233926 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233958 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.233984 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234011 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234083 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234115 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234172 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234200 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234291 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234325 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234348 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234375 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234400 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234424 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234454 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234482 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234510 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234536 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234568 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234593 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234612 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234630 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234649 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234666 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234686 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234703 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234720 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234738 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234757 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234778 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234799 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234819 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234839 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234863 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234887 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234913 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234938 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234962 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.234986 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235010 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235037 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235060 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235102 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235124 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235176 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235202 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235233 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235257 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235280 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235297 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235314 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235333 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235350 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235372 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235395 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235417 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235440 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235462 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235483 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235506 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235528 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235551 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235574 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235804 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235822 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235842 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235861 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235895 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235921 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235932 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.235944 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236026 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236057 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236089 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236118 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236328 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236361 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236392 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236461 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236490 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236524 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236550 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236577 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236581 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236626 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236651 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236652 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236698 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236730 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236756 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236785 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236808 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.236831 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237033 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237068 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237096 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237121 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237170 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237198 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237220 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237246 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237265 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237285 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237295 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237305 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237355 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237384 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237391 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237451 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237547 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237602 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237640 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237668 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237696 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237723 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237748 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237763 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237776 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237803 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237830 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237854 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237877 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237907 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237944 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237972 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237996 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238020 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238049 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238075 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238100 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238134 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238182 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238210 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238236 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238262 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238287 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238309 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238333 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238355 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238382 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238404 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238427 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238451 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238475 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238499 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238522 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238545 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238568 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238602 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238630 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238652 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238672 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238696 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238721 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238743 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238773 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238801 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238832 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238859 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238971 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-script-lib\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239006 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-netns\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239038 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-var-lib-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239062 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239089 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-systemd-units\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239117 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fvlx\" (UniqueName: \"kubernetes.io/projected/39cfa1ec-4912-4550-9ead-7f6113db2221-kube-api-access-9fvlx\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239251 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239291 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239320 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-system-cni-dir\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239352 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239415 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-cnibin\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239445 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-os-release\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239471 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cnibin\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239496 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2419d2de-214a-4a13-b941-2acd571f0360-proxy-tls\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239525 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2419d2de-214a-4a13-b941-2acd571f0360-mcd-auth-proxy-config\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239557 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239601 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239634 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-cni-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239667 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-conf-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239692 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-config\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239718 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-ovn\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239741 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-ovn-kubernetes\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239765 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-netd\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239789 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cni-binary-copy\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239819 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239849 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239872 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-os-release\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239900 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239926 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-node-log\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239948 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-log-socket\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239974 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-system-cni-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240000 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-daemon-config\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240024 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-env-overrides\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240049 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-hostroot\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240083 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240118 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-netns\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240170 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-kubelet\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240197 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/baeca99f-afc8-4129-8110-51e79dd8b054-ovn-node-metrics-cert\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240224 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240257 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnvfl\" (UniqueName: \"kubernetes.io/projected/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-kube-api-access-fnvfl\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240280 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-etc-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240313 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240341 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-slash\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240379 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-bin\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240404 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-etc-kubernetes\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240434 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240468 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-socket-dir-parent\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240494 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2419d2de-214a-4a13-b941-2acd571f0360-rootfs\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240517 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-systemd\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240551 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-multus-certs\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240578 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-cni-bin\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240609 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240638 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240680 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh4sp\" (UniqueName: \"kubernetes.io/projected/2419d2de-214a-4a13-b941-2acd571f0360-kube-api-access-rh4sp\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240710 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240738 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-kubelet\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240766 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-cni-multus\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240796 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvwpl\" (UniqueName: \"kubernetes.io/projected/027f048a-2237-4fa6-9bc5-4bd981c13b7a-kube-api-access-qvwpl\") pod \"node-resolver-fdhpt\" (UID: \"027f048a-2237-4fa6-9bc5-4bd981c13b7a\") " pod="openshift-dns/node-resolver-fdhpt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240825 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/39cfa1ec-4912-4550-9ead-7f6113db2221-cni-binary-copy\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240854 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/027f048a-2237-4fa6-9bc5-4bd981c13b7a-hosts-file\") pod \"node-resolver-fdhpt\" (UID: \"027f048a-2237-4fa6-9bc5-4bd981c13b7a\") " pod="openshift-dns/node-resolver-fdhpt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240890 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kckf4\" (UniqueName: \"kubernetes.io/projected/baeca99f-afc8-4129-8110-51e79dd8b054-kube-api-access-kckf4\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240916 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240945 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-k8s-cni-cncf-io\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240968 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241204 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241225 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241240 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241255 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241270 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241291 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241306 4788 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241320 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241334 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241351 4788 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241366 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241380 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241394 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241408 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241422 4788 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241556 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241593 4788 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241616 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241635 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241653 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241670 4788 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241687 4788 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241826 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241846 4788 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241861 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241880 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241896 4788 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241911 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241926 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.241942 4788 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244483 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.245774 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.247111 4788 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.248397 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250790 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.261372 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.262298 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237905 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.265242 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.265446 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.265475 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.265586 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.265795 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.237972 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238008 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238441 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238728 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238769 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238803 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.238944 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239019 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239214 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239532 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239528 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239617 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239644 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239983 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.239996 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240103 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240542 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240556 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.240592 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.242052 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.265947 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.266029 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.266167 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.242056 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.242354 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.242923 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.243417 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.243710 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.243738 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244004 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244447 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244465 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244548 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244607 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244765 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244844 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244934 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.244968 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.245191 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.245517 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.245532 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.245808 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.245912 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.245933 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246116 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246249 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246298 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246423 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246433 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246715 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246739 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246766 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.246986 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.247075 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.247160 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.247188 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.247549 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.247777 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250120 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250214 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250537 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250550 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250507 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.250745 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:45:18.750722287 +0000 UTC m=+21.200437835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250841 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250963 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250981 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.251025 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.251644 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.250621 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.251706 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.252007 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.252021 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.252392 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.252721 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.253236 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.253720 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.254111 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.254386 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.254648 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.254874 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.258777 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.259190 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.259207 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.259356 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.259639 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.259661 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.259781 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.260497 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.260917 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.261220 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.261422 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.261465 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.261556 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.261807 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.261806 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.262013 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.262317 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.266768 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.266790 4788 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.262454 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.262483 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.262548 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.262648 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.262664 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.262819 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.263087 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.263266 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.263160 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.263415 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.263782 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.264387 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.264426 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.264755 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.264828 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.264993 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.264408 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.266947 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.267058 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.267101 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.267687 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.269006 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.270289 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:18.770252991 +0000 UTC m=+21.219968539 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.271865 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.272192 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.272237 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.273530 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.273845 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.273989 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.274478 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.274695 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.274809 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.276959 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.277356 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.277559 4788 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.277895 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:18.77783872 +0000 UTC m=+21.227554278 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.277562 4788 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.277999 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.278049 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.278662 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.278663 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.279324 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.279557 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.279583 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:18.779253937 +0000 UTC m=+21.228969525 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.279725 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.279969 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.280301 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.280598 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.281261 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.281398 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.281957 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.282539 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.283826 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.284306 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.284471 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.284918 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.285041 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.285842 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.292792 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.295365 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.295600 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.295969 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.296156 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.296980 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.299808 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.299855 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.299871 4788 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.299854 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.299939 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:18.799918481 +0000 UTC m=+21.249634029 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.304721 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.305215 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.305388 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.305440 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.305533 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.308049 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.308019 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.309558 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.314947 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.322622 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.326403 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.359576 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360008 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2419d2de-214a-4a13-b941-2acd571f0360-proxy-tls\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360033 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2419d2de-214a-4a13-b941-2acd571f0360-mcd-auth-proxy-config\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360050 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360066 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-cni-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360082 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-conf-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360645 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cnibin\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360664 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-config\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360392 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-cni-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360680 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-ovn-kubernetes\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360719 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-netd\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360743 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-netd\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360799 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cni-binary-copy\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360903 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-ovn\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360935 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-os-release\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360965 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-node-log\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360990 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-log-socket\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361016 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-system-cni-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361046 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-daemon-config\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361059 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2419d2de-214a-4a13-b941-2acd571f0360-mcd-auth-proxy-config\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361088 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-env-overrides\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361119 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-netns\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361209 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-os-release\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360221 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361272 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cnibin\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361330 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-config\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361345 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-kubelet\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361373 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-ovn-kubernetes\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361374 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-hostroot\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361403 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361418 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-hostroot\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361903 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361932 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-cni-binary-copy\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.360512 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-conf-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.361422 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnvfl\" (UniqueName: \"kubernetes.io/projected/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-kube-api-access-fnvfl\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362113 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-ovn\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362154 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-kubelet\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362185 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-netns\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362221 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-log-socket\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362238 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-node-log\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362257 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-env-overrides\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362316 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-etc-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362320 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-system-cni-dir\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362350 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/baeca99f-afc8-4129-8110-51e79dd8b054-ovn-node-metrics-cert\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362506 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-slash\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362546 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-bin\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362577 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-slash\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362638 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-bin\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362723 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-socket-dir-parent\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362756 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-etc-kubernetes\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362772 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-systemd\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362797 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-etc-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362802 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-daemon-config\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362850 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-etc-kubernetes\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362864 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-multus-certs\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362885 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-systemd\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362898 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2419d2de-214a-4a13-b941-2acd571f0360-rootfs\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362916 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-multus-certs\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362928 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-multus-socket-dir-parent\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.362960 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2419d2de-214a-4a13-b941-2acd571f0360-rootfs\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363106 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-cni-bin\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363154 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363192 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-kubelet\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363216 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-cni-multus\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363240 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvwpl\" (UniqueName: \"kubernetes.io/projected/027f048a-2237-4fa6-9bc5-4bd981c13b7a-kube-api-access-qvwpl\") pod \"node-resolver-fdhpt\" (UID: \"027f048a-2237-4fa6-9bc5-4bd981c13b7a\") " pod="openshift-dns/node-resolver-fdhpt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363261 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh4sp\" (UniqueName: \"kubernetes.io/projected/2419d2de-214a-4a13-b941-2acd571f0360-kube-api-access-rh4sp\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363284 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/39cfa1ec-4912-4550-9ead-7f6113db2221-cni-binary-copy\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363306 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/027f048a-2237-4fa6-9bc5-4bd981c13b7a-hosts-file\") pod \"node-resolver-fdhpt\" (UID: \"027f048a-2237-4fa6-9bc5-4bd981c13b7a\") " pod="openshift-dns/node-resolver-fdhpt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363327 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363351 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-k8s-cni-cncf-io\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363373 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363396 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kckf4\" (UniqueName: \"kubernetes.io/projected/baeca99f-afc8-4129-8110-51e79dd8b054-kube-api-access-kckf4\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363417 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-netns\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363442 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-var-lib-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363467 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363481 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363493 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-script-lib\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.363750 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2419d2de-214a-4a13-b941-2acd571f0360-proxy-tls\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.364423 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-kubelet\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.364562 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.364888 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-script-lib\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.364938 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-netns\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.365079 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fvlx\" (UniqueName: \"kubernetes.io/projected/39cfa1ec-4912-4550-9ead-7f6113db2221-kube-api-access-9fvlx\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.365107 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-systemd-units\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.365160 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-cni-bin\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.366625 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-var-lib-cni-multus\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.366734 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-cnibin\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.366866 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-os-release\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.366889 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-systemd-units\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.366896 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-system-cni-dir\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.366942 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-var-lib-openvswitch\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.366984 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-cnibin\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367009 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-host-run-k8s-cni-cncf-io\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367017 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/baeca99f-afc8-4129-8110-51e79dd8b054-ovn-node-metrics-cert\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.366866 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367058 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/027f048a-2237-4fa6-9bc5-4bd981c13b7a-hosts-file\") pod \"node-resolver-fdhpt\" (UID: \"027f048a-2237-4fa6-9bc5-4bd981c13b7a\") " pod="openshift-dns/node-resolver-fdhpt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367066 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-system-cni-dir\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367069 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/39cfa1ec-4912-4550-9ead-7f6113db2221-os-release\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367208 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367352 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367372 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367382 4788 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367393 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367404 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367413 4788 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367423 4788 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367433 4788 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367443 4788 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367459 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367474 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367485 4788 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367495 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367504 4788 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367514 4788 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367522 4788 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367533 4788 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367546 4788 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367555 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367564 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367574 4788 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367585 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367594 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367605 4788 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367614 4788 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367624 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367633 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367642 4788 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367652 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367662 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367672 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367682 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367692 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367702 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367711 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367721 4788 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367732 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367744 4788 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367755 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367765 4788 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367776 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367787 4788 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367796 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367807 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367817 4788 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367828 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367839 4788 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367850 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367858 4788 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367868 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367878 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367887 4788 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367897 4788 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367906 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367918 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367927 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367929 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/39cfa1ec-4912-4550-9ead-7f6113db2221-cni-binary-copy\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367941 4788 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.367991 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369768 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369792 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369809 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369824 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369837 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369944 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369958 4788 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369970 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369982 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.369997 4788 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370009 4788 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370021 4788 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370033 4788 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370044 4788 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370057 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370069 4788 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370082 4788 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370095 4788 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370119 4788 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370132 4788 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370167 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370180 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370192 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370205 4788 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370221 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370234 4788 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.370246 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372689 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372703 4788 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372717 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372727 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372739 4788 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372758 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372771 4788 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372782 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372793 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372809 4788 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372819 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372830 4788 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372841 4788 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372855 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372874 4788 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372891 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372904 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372918 4788 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372930 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372940 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372949 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372960 4788 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372972 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372985 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.372998 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373011 4788 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373032 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373046 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373062 4788 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373074 4788 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373089 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373105 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373117 4788 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373127 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373152 4788 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373166 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373179 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373192 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373203 4788 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373214 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373227 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373238 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373248 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373257 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373334 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373346 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373367 4788 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373376 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373385 4788 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373395 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373404 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373413 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373423 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373434 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373443 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373453 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373462 4788 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373471 4788 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373482 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373492 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373501 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373510 4788 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373519 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373527 4788 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373536 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373546 4788 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373582 4788 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373591 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373603 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373615 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373624 4788 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373633 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373644 4788 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373653 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373661 4788 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373671 4788 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373680 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373689 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.373699 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.379995 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.386266 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.388577 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvwpl\" (UniqueName: \"kubernetes.io/projected/027f048a-2237-4fa6-9bc5-4bd981c13b7a-kube-api-access-qvwpl\") pod \"node-resolver-fdhpt\" (UID: \"027f048a-2237-4fa6-9bc5-4bd981c13b7a\") " pod="openshift-dns/node-resolver-fdhpt" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.389466 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.390633 4788 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.390713 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.390835 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh4sp\" (UniqueName: \"kubernetes.io/projected/2419d2de-214a-4a13-b941-2acd571f0360-kube-api-access-rh4sp\") pod \"machine-config-daemon-dxfdf\" (UID: \"2419d2de-214a-4a13-b941-2acd571f0360\") " pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.392349 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnvfl\" (UniqueName: \"kubernetes.io/projected/6c8cb468-2b03-45c9-93e9-abbcc4c075a3-kube-api-access-fnvfl\") pod \"multus-additional-cni-plugins-szmv7\" (UID: \"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\") " pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.392656 4788 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63" exitCode=255 Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.392697 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63"} Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.393230 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kckf4\" (UniqueName: \"kubernetes.io/projected/baeca99f-afc8-4129-8110-51e79dd8b054-kube-api-access-kckf4\") pod \"ovnkube-node-cwgns\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.395054 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.395370 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fvlx\" (UniqueName: \"kubernetes.io/projected/39cfa1ec-4912-4550-9ead-7f6113db2221-kube-api-access-9fvlx\") pod \"multus-rzgfk\" (UID: \"39cfa1ec-4912-4550-9ead-7f6113db2221\") " pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.396007 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.406514 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.414982 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.417156 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.424229 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.433692 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.443515 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.460900 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.472746 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.483919 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 10 14:45:18 crc kubenswrapper[4788]: W1010 14:45:18.499489 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-d99b32b895ba62ec6a39153d5d5e49543c1c3b2778b72eb59dff786ac471477e WatchSource:0}: Error finding container d99b32b895ba62ec6a39153d5d5e49543c1c3b2778b72eb59dff786ac471477e: Status 404 returned error can't find the container with id d99b32b895ba62ec6a39153d5d5e49543c1c3b2778b72eb59dff786ac471477e Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.500213 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.503038 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.517408 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.521476 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rzgfk" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.530345 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-szmv7" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.543528 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.543702 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.562412 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.579997 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.587013 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.599192 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.610321 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fdhpt" Oct 10 14:45:18 crc kubenswrapper[4788]: W1010 14:45:18.614169 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c8cb468_2b03_45c9_93e9_abbcc4c075a3.slice/crio-6b017a1d37e842e2da6e723fb9813826b013cf8ff52cde5d5657b8fea2bfa5c9 WatchSource:0}: Error finding container 6b017a1d37e842e2da6e723fb9813826b013cf8ff52cde5d5657b8fea2bfa5c9: Status 404 returned error can't find the container with id 6b017a1d37e842e2da6e723fb9813826b013cf8ff52cde5d5657b8fea2bfa5c9 Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.620424 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.627176 4788 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.627233 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.635306 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.647985 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.657997 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.682716 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.695604 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.714995 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.736431 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.748516 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.760069 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.780633 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.782491 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.782618 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.782671 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.782711 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.782904 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.782936 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.782956 4788 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.783022 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:19.783001205 +0000 UTC m=+22.232716773 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.783104 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:45:19.783095977 +0000 UTC m=+22.232811535 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.783190 4788 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.783222 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:19.78321357 +0000 UTC m=+22.232929138 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.783289 4788 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.783324 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:19.783316133 +0000 UTC m=+22.233031681 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.797477 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.821555 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.835263 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:18 crc kubenswrapper[4788]: I1010 14:45:18.883717 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.883871 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.883887 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.883898 4788 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:18 crc kubenswrapper[4788]: E1010 14:45:18.883954 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:19.883937271 +0000 UTC m=+22.333652819 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.398895 4788 generic.go:334] "Generic (PLEG): container finished" podID="6c8cb468-2b03-45c9-93e9-abbcc4c075a3" containerID="9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182" exitCode=0 Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.398980 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerDied","Data":"9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.399067 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerStarted","Data":"6b017a1d37e842e2da6e723fb9813826b013cf8ff52cde5d5657b8fea2bfa5c9"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.402066 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.402121 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.402152 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6fc025242b870fde2c31aff9f007704c0eca7c3954f97c127eab7715e7b700ed"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.403819 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.403872 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d99b32b895ba62ec6a39153d5d5e49543c1c3b2778b72eb59dff786ac471477e"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.405460 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e2097a224696992a6497f1fb323182e76800f5bdaa13a6eee39eefe752f5f668"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.407076 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rzgfk" event={"ID":"39cfa1ec-4912-4550-9ead-7f6113db2221","Type":"ContainerStarted","Data":"7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.407232 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rzgfk" event={"ID":"39cfa1ec-4912-4550-9ead-7f6113db2221","Type":"ContainerStarted","Data":"dc1236cb47959583ee8696aa009f8736978c62b836d7e5d2695c35350f4c13ee"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.409723 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fdhpt" event={"ID":"027f048a-2237-4fa6-9bc5-4bd981c13b7a","Type":"ContainerStarted","Data":"f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.409756 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fdhpt" event={"ID":"027f048a-2237-4fa6-9bc5-4bd981c13b7a","Type":"ContainerStarted","Data":"106c68a025fc9e4f6f8a05a93dec249aa42a5a0af78795764571df6930ab1bf8"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.411830 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.411859 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.411869 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"60ca8b42ac0816a375c5df221b7b4769555d37cc7aae048ffae1117f84d87b92"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.413198 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717" exitCode=0 Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.413736 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.413760 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"c87fdf5d11a77f4ff6ffdb2328a38057d29f20709779bd6c141fd8dc7502e076"} Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.414362 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.422382 4788 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.422672 4788 scope.go:117] "RemoveContainer" containerID="435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.439674 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.454215 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.489595 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.524824 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.552100 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.575249 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.596603 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.615436 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.630220 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.653882 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.668988 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.682875 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.697119 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.715054 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.729787 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.750943 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.765737 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.790829 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.792891 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.793108 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.793178 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:45:21.79312049 +0000 UTC m=+24.242836038 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.793312 4788 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.793419 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:21.793395187 +0000 UTC m=+24.243110895 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.793491 4788 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.793590 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:21.793575961 +0000 UTC m=+24.243291509 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.793324 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.793721 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.794190 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.794229 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.794249 4788 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.794309 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:21.79429803 +0000 UTC m=+24.244013578 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.808885 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.823016 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.837217 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.851280 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.865378 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:19Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:19 crc kubenswrapper[4788]: I1010 14:45:19.895179 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.895430 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.895471 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.895484 4788 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:19 crc kubenswrapper[4788]: E1010 14:45:19.895555 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:21.895533085 +0000 UTC m=+24.345248843 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.233201 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:20 crc kubenswrapper[4788]: E1010 14:45:20.233680 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.233328 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.233301 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:20 crc kubenswrapper[4788]: E1010 14:45:20.233789 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:20 crc kubenswrapper[4788]: E1010 14:45:20.233963 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.238983 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.239966 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.290111 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.291623 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.351831 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.353093 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.354276 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.361460 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.362555 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.363498 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.364496 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.365897 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.367050 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.421924 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.441553 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.445773 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.446846 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.447502 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.448171 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.448622 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.450224 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.450902 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.451473 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.452564 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.453034 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.454185 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.454729 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.455529 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.455835 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.458669 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.459215 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.460624 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.461196 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.461697 4788 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.461817 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.464322 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.464879 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.465524 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.467160 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.468347 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.468929 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.469873 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.470681 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.471579 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.472199 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.473426 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.473783 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.474109 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.474999 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.475992 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.477104 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.478667 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.479198 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.480191 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.480671 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.480728 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.480750 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.480764 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.480774 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.480784 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95"} Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.480796 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerStarted","Data":"6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d"} Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.486127 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.502593 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.514606 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.530100 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.549934 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.565009 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.567218 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.578590 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.588591 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.599679 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.614611 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.629983 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.630395 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.643122 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.663102 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.676692 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.695496 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.714951 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.737236 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.751168 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.765902 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.778799 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.795301 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:20 crc kubenswrapper[4788]: I1010 14:45:20.813830 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:20Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.435522 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.435947 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.436977 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0"} Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.438934 4788 generic.go:334] "Generic (PLEG): container finished" podID="6c8cb468-2b03-45c9-93e9-abbcc4c075a3" containerID="6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d" exitCode=0 Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.439034 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerDied","Data":"6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d"} Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.451727 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.465431 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.478675 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.490801 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.503007 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.522959 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.535036 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.548837 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.561418 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.577424 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.593054 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.623172 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.636093 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.648799 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.662163 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.673799 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.690452 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.704281 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.722630 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.733687 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.748684 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.764097 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.776877 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.798262 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.812998 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.814431 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.814614 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:45:25.8145869 +0000 UTC m=+28.264302478 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.814687 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.814754 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.814825 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.814844 4788 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.814897 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:25.814882198 +0000 UTC m=+28.264597746 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.814903 4788 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.814964 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:25.8149511 +0000 UTC m=+28.264666678 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.815006 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.815040 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.815064 4788 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.815186 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:25.815126624 +0000 UTC m=+28.264842212 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.830413 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:21Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:21 crc kubenswrapper[4788]: I1010 14:45:21.916249 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.916499 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.916526 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.916541 4788 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:21 crc kubenswrapper[4788]: E1010 14:45:21.916622 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:25.916598624 +0000 UTC m=+28.366314172 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.021369 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.024614 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.029792 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.039178 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.051703 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.070867 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.081208 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.092238 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.110621 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.125068 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.140485 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.152985 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.173998 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.189396 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.205798 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.220071 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.231038 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.232908 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.232971 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.232991 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:22 crc kubenswrapper[4788]: E1010 14:45:22.233064 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:22 crc kubenswrapper[4788]: E1010 14:45:22.233245 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:22 crc kubenswrapper[4788]: E1010 14:45:22.233496 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.246863 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.257202 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.269041 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.281596 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.299056 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.329344 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.371468 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.410980 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.444625 4788 generic.go:334] "Generic (PLEG): container finished" podID="6c8cb468-2b03-45c9-93e9-abbcc4c075a3" containerID="72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9" exitCode=0 Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.444722 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerDied","Data":"72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9"} Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.453889 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.492452 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.530411 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.577844 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.613599 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.652984 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.693002 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.730158 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.771372 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.781368 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2pg5g"] Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.781726 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.802957 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.822751 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.825967 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klhwj\" (UniqueName: \"kubernetes.io/projected/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-kube-api-access-klhwj\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.826086 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-host\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.826125 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-serviceca\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.843205 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.863498 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.891441 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.927369 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klhwj\" (UniqueName: \"kubernetes.io/projected/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-kube-api-access-klhwj\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.927457 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-host\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.927495 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-serviceca\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.927615 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-host\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.928423 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-serviceca\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.942293 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.961224 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klhwj\" (UniqueName: \"kubernetes.io/projected/5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d-kube-api-access-klhwj\") pod \"node-ca-2pg5g\" (UID: \"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\") " pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:22 crc kubenswrapper[4788]: I1010 14:45:22.992102 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:22Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.031407 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.082046 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.093657 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2pg5g" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.114562 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.156683 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.191273 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.237479 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.269957 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.311394 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.354847 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.392304 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.430389 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.461258 4788 generic.go:334] "Generic (PLEG): container finished" podID="6c8cb468-2b03-45c9-93e9-abbcc4c075a3" containerID="171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb" exitCode=0 Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.461319 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerDied","Data":"171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb"} Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.464171 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2pg5g" event={"ID":"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d","Type":"ContainerStarted","Data":"0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135"} Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.464248 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2pg5g" event={"ID":"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d","Type":"ContainerStarted","Data":"ba758554e63394128f09ba5639d85cb02e096653c9b3719572713b52f8aa51d8"} Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.471957 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.473489 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.518741 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.551521 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.593255 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.634849 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.673658 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.715340 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.725180 4788 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.727801 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.727835 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.727845 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.727949 4788 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.753159 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.803890 4788 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.804191 4788 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.805534 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.805564 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.805572 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.805586 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.805596 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:23Z","lastTransitionTime":"2025-10-10T14:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:23 crc kubenswrapper[4788]: E1010 14:45:23.818111 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.823199 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.823232 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.823242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.823259 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.823268 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:23Z","lastTransitionTime":"2025-10-10T14:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:23 crc kubenswrapper[4788]: E1010 14:45:23.834179 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.838086 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.838123 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.838133 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.838169 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.838181 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:23Z","lastTransitionTime":"2025-10-10T14:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.839002 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: E1010 14:45:23.858310 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.862876 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.863167 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.863193 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.863227 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.863248 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:23Z","lastTransitionTime":"2025-10-10T14:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.874409 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: E1010 14:45:23.877956 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.882118 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.882193 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.882215 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.882242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.882263 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:23Z","lastTransitionTime":"2025-10-10T14:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:23 crc kubenswrapper[4788]: E1010 14:45:23.895765 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: E1010 14:45:23.895888 4788 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.897724 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.897771 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.897787 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.897810 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.897827 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:23Z","lastTransitionTime":"2025-10-10T14:45:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.911959 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.953361 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:23 crc kubenswrapper[4788]: I1010 14:45:23.996649 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:23Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.000531 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.000579 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.000588 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.000604 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.000615 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.032656 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.073510 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.103120 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.103190 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.103205 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.103226 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.103237 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.120079 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.159216 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.191267 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.209122 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.209188 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.209200 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.209216 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.209229 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.233256 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.233302 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.233388 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:24 crc kubenswrapper[4788]: E1010 14:45:24.233411 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.233398 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: E1010 14:45:24.233615 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:24 crc kubenswrapper[4788]: E1010 14:45:24.233810 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.273396 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.310897 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.311733 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.311769 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.311778 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.311794 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.311803 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.360530 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.403255 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.414244 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.414300 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.414312 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.414327 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.414336 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.440687 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.470977 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.482183 4788 generic.go:334] "Generic (PLEG): container finished" podID="6c8cb468-2b03-45c9-93e9-abbcc4c075a3" containerID="6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361" exitCode=0 Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.482227 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerDied","Data":"6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.510387 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.516932 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.516960 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.516969 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.516982 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.516992 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.558610 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.590652 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.619657 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.619689 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.619698 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.619712 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.619720 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.634654 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.672854 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.714529 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.722597 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.722652 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.722663 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.722697 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.722709 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.754269 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.791583 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.825092 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.825124 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.825132 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.825162 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.825171 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.831474 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.871447 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.910046 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.927338 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.927367 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.927378 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.927392 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.927401 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:24Z","lastTransitionTime":"2025-10-10T14:45:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.956643 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:24 crc kubenswrapper[4788]: I1010 14:45:24.992368 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:24Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.029997 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.030042 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.030054 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.030071 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.030083 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.037989 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.073197 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.113165 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.132704 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.132736 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.132749 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.132773 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.132785 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.240226 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.240273 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.240286 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.240304 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.240317 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.344002 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.344051 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.344067 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.344086 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.344098 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.447222 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.447306 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.447324 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.447463 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.447485 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.489207 4788 generic.go:334] "Generic (PLEG): container finished" podID="6c8cb468-2b03-45c9-93e9-abbcc4c075a3" containerID="65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7" exitCode=0 Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.489274 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerDied","Data":"65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.495503 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.496034 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.496122 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.496398 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.503274 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.521779 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.522001 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.523521 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.538915 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.550402 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.550521 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.550612 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.550700 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.550781 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.552884 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.570133 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.580714 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.594757 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.606921 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.618875 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.632715 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.648954 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.652925 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.652962 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.652970 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.652983 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.652992 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.664214 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.676970 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.691160 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.721098 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.755392 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.755428 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.755436 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.755453 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.755466 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.758517 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.789964 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.829807 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857235 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857327 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857355 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857367 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857385 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857396 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857442 4788 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857487 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:33.85747526 +0000 UTC m=+36.307190808 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857362 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857525 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:45:33.857504221 +0000 UTC m=+36.307219819 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857580 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.857624 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857743 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857768 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857781 4788 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857818 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:33.857809109 +0000 UTC m=+36.307524657 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857880 4788 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.857914 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:33.857905391 +0000 UTC m=+36.307621029 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.870653 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.918281 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.951800 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.958357 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.958527 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.958551 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.958562 4788 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:25 crc kubenswrapper[4788]: E1010 14:45:25.958617 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:33.958603942 +0000 UTC m=+36.408319490 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.960394 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.960431 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.960446 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.960467 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.960483 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:25Z","lastTransitionTime":"2025-10-10T14:45:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:25 crc kubenswrapper[4788]: I1010 14:45:25.997224 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:25Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.038322 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.062874 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.062918 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.062932 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.062952 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.062968 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.075417 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.113989 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.151734 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.164684 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.164720 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.164730 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.164745 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.164757 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.194422 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.233625 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.233671 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.233708 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:26 crc kubenswrapper[4788]: E1010 14:45:26.233747 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:26 crc kubenswrapper[4788]: E1010 14:45:26.233816 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:26 crc kubenswrapper[4788]: E1010 14:45:26.233884 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.234778 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.267338 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.267374 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.267383 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.267397 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.267407 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.271855 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.310394 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.370082 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.370654 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.370884 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.371052 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.371242 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.474216 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.474254 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.474267 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.474285 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.474297 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.502944 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" event={"ID":"6c8cb468-2b03-45c9-93e9-abbcc4c075a3","Type":"ContainerStarted","Data":"91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.516334 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.536551 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.549062 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.565801 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.577180 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.577344 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.577540 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.577728 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.577818 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.577773 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.594910 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.606481 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.635444 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.675984 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.681011 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.681057 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.681069 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.681089 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.681102 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.716431 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.752987 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.791746 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.800873 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.800954 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.800971 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.800988 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.801004 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.838314 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.876282 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.903560 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.903891 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.904012 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.904097 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.906868 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:26Z","lastTransitionTime":"2025-10-10T14:45:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:26 crc kubenswrapper[4788]: I1010 14:45:26.913167 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.010012 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.010080 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.010095 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.010119 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.010155 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.113035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.113105 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.113121 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.113498 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.113525 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.217060 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.217256 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.217287 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.217327 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.217355 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.320515 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.320856 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.320927 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.321006 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.321066 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.423284 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.423332 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.423535 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.423568 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.423578 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.526767 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.526840 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.526858 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.526892 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.526909 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.629166 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.629204 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.629213 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.629228 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.629237 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.731393 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.731446 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.731456 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.731471 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.731480 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.833724 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.833778 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.833795 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.833817 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.833834 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.935654 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.935695 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.935705 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.935720 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:27 crc kubenswrapper[4788]: I1010 14:45:27.935729 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:27Z","lastTransitionTime":"2025-10-10T14:45:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.038526 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.038579 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.038594 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.038615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.038627 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.141026 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.141069 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.141082 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.141103 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.141114 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.232884 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:28 crc kubenswrapper[4788]: E1010 14:45:28.233009 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.233265 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:28 crc kubenswrapper[4788]: E1010 14:45:28.233406 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.233265 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:28 crc kubenswrapper[4788]: E1010 14:45:28.233505 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.243449 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.243492 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.243502 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.243521 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.243534 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.252601 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.264406 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.274356 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.285394 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.297742 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.307703 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.325268 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.336613 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.345695 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.345729 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.345738 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.345751 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.345763 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.359621 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.374977 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.393017 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.407227 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.423958 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.438885 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.448182 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.448220 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.448230 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.448247 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.448259 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.452250 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.514026 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/0.log" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.518049 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72" exitCode=1 Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.518166 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.519436 4788 scope.go:117] "RemoveContainer" containerID="c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.534851 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.553835 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.554258 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.554276 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.555013 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.555035 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.556579 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.571333 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.588069 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.603158 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.621850 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.645625 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.658507 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.658588 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.658609 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.658643 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.658664 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.662164 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.676982 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.695447 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.709440 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.731800 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.746265 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.761324 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.761373 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.761387 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.761405 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.761419 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.770619 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:28Z\\\",\\\"message\\\":\\\"81 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134264 6081 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134883 6081 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 14:45:28.134905 6081 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 14:45:28.134949 6081 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 14:45:28.134961 6081 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 14:45:28.135113 6081 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 14:45:28.135150 6081 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 14:45:28.135205 6081 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 14:45:28.135231 6081 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 14:45:28.135239 6081 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 14:45:28.135251 6081 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 14:45:28.135270 6081 factory.go:656] Stopping watch factory\\\\nI1010 14:45:28.135289 6081 ovnkube.go:599] Stopped ovnkube\\\\nI1010 14:45:28.135305 6081 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.784646 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.863988 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.864036 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.864048 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.864067 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.864079 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.966968 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.967034 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.967046 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.967074 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:28 crc kubenswrapper[4788]: I1010 14:45:28.967090 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:28Z","lastTransitionTime":"2025-10-10T14:45:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.070541 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.070643 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.070666 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.070698 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.070721 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.173556 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.173634 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.173652 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.173691 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.173707 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.276029 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.276063 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.276072 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.276086 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.276095 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.379043 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.379128 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.379173 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.379204 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.379225 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.481885 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.481965 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.481985 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.482016 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.482036 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.529817 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/0.log" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.532985 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.533382 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.549668 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.563588 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.584880 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.584974 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.584996 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.585028 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.585052 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.588061 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.607682 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.625560 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.644221 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.664283 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:28Z\\\",\\\"message\\\":\\\"81 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134264 6081 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134883 6081 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 14:45:28.134905 6081 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 14:45:28.134949 6081 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 14:45:28.134961 6081 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 14:45:28.135113 6081 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 14:45:28.135150 6081 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 14:45:28.135205 6081 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 14:45:28.135231 6081 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 14:45:28.135239 6081 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 14:45:28.135251 6081 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 14:45:28.135270 6081 factory.go:656] Stopping watch factory\\\\nI1010 14:45:28.135289 6081 ovnkube.go:599] Stopped ovnkube\\\\nI1010 14:45:28.135305 6081 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.680195 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.688176 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.688204 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.688213 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.688227 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.688236 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.694280 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.707906 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.719584 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.731491 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.749844 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.764319 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.776435 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.790509 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.790578 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.790588 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.790604 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.790615 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.894206 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.894277 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.894296 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.894371 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.894399 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.997573 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.997616 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.997629 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.997647 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:29 crc kubenswrapper[4788]: I1010 14:45:29.997657 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:29Z","lastTransitionTime":"2025-10-10T14:45:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.102124 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.102219 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.102233 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.102255 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.102267 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.205541 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.205619 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.205632 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.205686 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.205699 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.233301 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.233346 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.233353 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:30 crc kubenswrapper[4788]: E1010 14:45:30.233473 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:30 crc kubenswrapper[4788]: E1010 14:45:30.233738 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:30 crc kubenswrapper[4788]: E1010 14:45:30.233858 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.307535 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.307615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.307626 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.307638 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.307648 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.410231 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.410271 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.410283 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.410298 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.410309 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.513406 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.513447 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.513457 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.513473 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.513483 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.538744 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/1.log" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.539584 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/0.log" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.543261 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0" exitCode=1 Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.543478 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.543647 4788 scope.go:117] "RemoveContainer" containerID="c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.544944 4788 scope.go:117] "RemoveContainer" containerID="5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0" Oct 10 14:45:30 crc kubenswrapper[4788]: E1010 14:45:30.545255 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.565326 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.578934 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.594990 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.604786 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.615660 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.615697 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.615706 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.615720 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.615732 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.617702 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.628850 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.649400 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:28Z\\\",\\\"message\\\":\\\"81 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134264 6081 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134883 6081 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 14:45:28.134905 6081 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 14:45:28.134949 6081 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 14:45:28.134961 6081 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 14:45:28.135113 6081 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 14:45:28.135150 6081 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 14:45:28.135205 6081 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 14:45:28.135231 6081 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 14:45:28.135239 6081 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 14:45:28.135251 6081 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 14:45:28.135270 6081 factory.go:656] Stopping watch factory\\\\nI1010 14:45:28.135289 6081 ovnkube.go:599] Stopped ovnkube\\\\nI1010 14:45:28.135305 6081 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.670745 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.692357 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.714347 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.718889 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.718920 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.718929 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.718947 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.718963 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.730794 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.750902 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.776400 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.791824 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.804643 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:30Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.821878 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.821908 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.821916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.821929 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.821938 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.925084 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.925120 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.925129 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.925156 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:30 crc kubenswrapper[4788]: I1010 14:45:30.925166 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:30Z","lastTransitionTime":"2025-10-10T14:45:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.027111 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.027480 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.027558 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.027629 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.027710 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.053013 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx"] Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.053948 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.056892 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.056919 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.090083 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.106298 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.121225 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.131647 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.131715 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.131730 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.131753 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.131766 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.136757 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.146288 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fff2dd72-c73c-455d-baa0-5f71322f1c0e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.146384 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fff2dd72-c73c-455d-baa0-5f71322f1c0e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.146426 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g58j\" (UniqueName: \"kubernetes.io/projected/fff2dd72-c73c-455d-baa0-5f71322f1c0e-kube-api-access-5g58j\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.146489 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fff2dd72-c73c-455d-baa0-5f71322f1c0e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.156569 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.170989 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.188116 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.208264 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.225542 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.235662 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.235730 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.235747 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.235770 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.235783 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.247172 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fff2dd72-c73c-455d-baa0-5f71322f1c0e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.247222 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g58j\" (UniqueName: \"kubernetes.io/projected/fff2dd72-c73c-455d-baa0-5f71322f1c0e-kube-api-access-5g58j\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.247246 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fff2dd72-c73c-455d-baa0-5f71322f1c0e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.247263 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fff2dd72-c73c-455d-baa0-5f71322f1c0e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.247997 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fff2dd72-c73c-455d-baa0-5f71322f1c0e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.248133 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fff2dd72-c73c-455d-baa0-5f71322f1c0e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.249988 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:28Z\\\",\\\"message\\\":\\\"81 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134264 6081 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134883 6081 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 14:45:28.134905 6081 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 14:45:28.134949 6081 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 14:45:28.134961 6081 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 14:45:28.135113 6081 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 14:45:28.135150 6081 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 14:45:28.135205 6081 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 14:45:28.135231 6081 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 14:45:28.135239 6081 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 14:45:28.135251 6081 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 14:45:28.135270 6081 factory.go:656] Stopping watch factory\\\\nI1010 14:45:28.135289 6081 ovnkube.go:599] Stopped ovnkube\\\\nI1010 14:45:28.135305 6081 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.253342 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fff2dd72-c73c-455d-baa0-5f71322f1c0e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.263158 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g58j\" (UniqueName: \"kubernetes.io/projected/fff2dd72-c73c-455d-baa0-5f71322f1c0e-kube-api-access-5g58j\") pod \"ovnkube-control-plane-749d76644c-448bx\" (UID: \"fff2dd72-c73c-455d-baa0-5f71322f1c0e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.265803 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.283038 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.301181 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.318861 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.335405 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.339039 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.339107 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.339119 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.339180 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.339195 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.352069 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.371440 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.442559 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.442601 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.442615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.442631 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.442640 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.467889 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.488654 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.504218 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.522246 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.537311 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.550865 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.550906 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.550916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.550931 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.550945 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.552500 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.554460 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" event={"ID":"fff2dd72-c73c-455d-baa0-5f71322f1c0e","Type":"ContainerStarted","Data":"ddbf4a4d3b79a0ddc2005cad568191fa989090ef91cb1fb16dc34409310d63ea"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.556364 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/1.log" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.560301 4788 scope.go:117] "RemoveContainer" containerID="5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0" Oct 10 14:45:31 crc kubenswrapper[4788]: E1010 14:45:31.560441 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.586646 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.603887 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.618330 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.633452 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.650260 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.653187 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.653242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.653255 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.653274 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.653287 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.664272 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.679640 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.692443 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.705381 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.716778 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.734132 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7ddea452ba0cb5d63b14c777f23437a55671e4bba282c8cf0a7676c0d4bfc72\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:28Z\\\",\\\"message\\\":\\\"81 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134264 6081 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1010 14:45:28.134883 6081 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1010 14:45:28.134905 6081 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1010 14:45:28.134949 6081 handler.go:208] Removed *v1.Node event handler 7\\\\nI1010 14:45:28.134961 6081 handler.go:208] Removed *v1.Node event handler 2\\\\nI1010 14:45:28.135113 6081 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1010 14:45:28.135150 6081 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1010 14:45:28.135205 6081 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1010 14:45:28.135231 6081 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1010 14:45:28.135239 6081 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1010 14:45:28.135251 6081 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1010 14:45:28.135270 6081 factory.go:656] Stopping watch factory\\\\nI1010 14:45:28.135289 6081 ovnkube.go:599] Stopped ovnkube\\\\nI1010 14:45:28.135305 6081 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1010 14\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.755193 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.755228 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.755239 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.755254 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.755263 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.757845 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.775919 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.787511 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.802416 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.814701 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-jqlwd"] Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.815160 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:31 crc kubenswrapper[4788]: E1010 14:45:31.815225 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.819839 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.831986 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.842425 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.855891 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.857394 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.857440 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.857452 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.857469 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.857482 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.868500 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.888048 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.899891 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.911568 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.923391 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.939660 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.953104 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.955360 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxgs4\" (UniqueName: \"kubernetes.io/projected/3d61e9aa-0688-436b-a8af-d6d8353c4350-kube-api-access-jxgs4\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.955415 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.959821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.959871 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.959883 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.959905 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.959916 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:31Z","lastTransitionTime":"2025-10-10T14:45:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.968846 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.988173 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:31 crc kubenswrapper[4788]: I1010 14:45:31.999734 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:31Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.009219 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.020527 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.037280 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.047418 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.056687 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.056753 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxgs4\" (UniqueName: \"kubernetes.io/projected/3d61e9aa-0688-436b-a8af-d6d8353c4350-kube-api-access-jxgs4\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:32 crc kubenswrapper[4788]: E1010 14:45:32.056887 4788 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:32 crc kubenswrapper[4788]: E1010 14:45:32.056963 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs podName:3d61e9aa-0688-436b-a8af-d6d8353c4350 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:32.556945367 +0000 UTC m=+35.006660915 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs") pod "network-metrics-daemon-jqlwd" (UID: "3d61e9aa-0688-436b-a8af-d6d8353c4350") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.058072 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.061545 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.061571 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.061579 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.061593 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.061603 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.068930 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.079786 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxgs4\" (UniqueName: \"kubernetes.io/projected/3d61e9aa-0688-436b-a8af-d6d8353c4350-kube-api-access-jxgs4\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.085214 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.097694 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.114569 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.124607 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.137492 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.149041 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.160252 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.163778 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.163839 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.163858 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.163882 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.163899 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.171490 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.182102 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.233077 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.233104 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.233077 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:32 crc kubenswrapper[4788]: E1010 14:45:32.233218 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:32 crc kubenswrapper[4788]: E1010 14:45:32.233351 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:32 crc kubenswrapper[4788]: E1010 14:45:32.233509 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.266123 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.266355 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.266417 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.266480 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.266571 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.368585 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.368619 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.368628 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.368641 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.368650 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.471719 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.471947 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.472028 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.472181 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.472274 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.561365 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:32 crc kubenswrapper[4788]: E1010 14:45:32.561671 4788 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:32 crc kubenswrapper[4788]: E1010 14:45:32.561748 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs podName:3d61e9aa-0688-436b-a8af-d6d8353c4350 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:33.561732042 +0000 UTC m=+36.011447590 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs") pod "network-metrics-daemon-jqlwd" (UID: "3d61e9aa-0688-436b-a8af-d6d8353c4350") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.564592 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" event={"ID":"fff2dd72-c73c-455d-baa0-5f71322f1c0e","Type":"ContainerStarted","Data":"761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.564630 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" event={"ID":"fff2dd72-c73c-455d-baa0-5f71322f1c0e","Type":"ContainerStarted","Data":"789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.574861 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.574898 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.574909 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.574927 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.574937 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.582987 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.595003 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.605274 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.615892 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.629303 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.639063 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.647799 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.656902 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.671992 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.677204 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.677250 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.677264 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.677281 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.677294 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.684381 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.706028 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.715759 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.728215 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.740558 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.757033 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.769912 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.780408 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.780446 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.780458 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.780498 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.780511 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.785740 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:32Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.883320 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.883358 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.883366 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.883382 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.883390 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.985626 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.985684 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.985703 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.985730 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:32 crc kubenswrapper[4788]: I1010 14:45:32.985748 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:32Z","lastTransitionTime":"2025-10-10T14:45:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.088713 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.088982 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.088998 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.089021 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.089036 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.190790 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.190817 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.190824 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.190836 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.190846 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.233214 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.233351 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.293617 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.293693 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.293712 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.293739 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.293761 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.396659 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.396990 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.397102 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.397218 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.397303 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.501076 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.501409 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.501485 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.501554 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.501617 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.573088 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.573348 4788 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.573643 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs podName:3d61e9aa-0688-436b-a8af-d6d8353c4350 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:35.573621913 +0000 UTC m=+38.023337481 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs") pod "network-metrics-daemon-jqlwd" (UID: "3d61e9aa-0688-436b-a8af-d6d8353c4350") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.605397 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.605480 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.605500 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.605539 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.605559 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.708668 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.708729 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.708740 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.708762 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.708779 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.812541 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.812583 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.812593 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.812614 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.812626 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.877504 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.877636 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.877702 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.877735 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:45:49.877705756 +0000 UTC m=+52.327421294 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.877826 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.877845 4788 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.877919 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:49.877899911 +0000 UTC m=+52.327615449 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.877997 4788 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.878017 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.878086 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.878110 4788 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.878042 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:49.878035314 +0000 UTC m=+52.327750862 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.878321 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:49.87827418 +0000 UTC m=+52.327989728 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.915127 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.915188 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.915197 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.915216 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.915226 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:33Z","lastTransitionTime":"2025-10-10T14:45:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:33 crc kubenswrapper[4788]: I1010 14:45:33.978771 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.978959 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.978980 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.978992 4788 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:33 crc kubenswrapper[4788]: E1010 14:45:33.979050 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:49.979034162 +0000 UTC m=+52.428749710 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.018492 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.018538 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.018549 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.018566 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.018576 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.121197 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.121242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.121253 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.121269 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.121281 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.223949 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.223994 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.224006 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.224023 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.224037 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.232924 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.232927 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.233071 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.233171 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.233373 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.233610 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.252070 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.252326 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.252414 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.252496 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.252746 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.274424 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:34Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.278733 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.278778 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.278789 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.278805 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.278815 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.291584 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:34Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.295608 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.295687 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.295705 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.295729 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.295748 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.307970 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:34Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.311736 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.311884 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.311966 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.312048 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.312122 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.323912 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:34Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.328031 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.328168 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.328252 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.328327 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.328391 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.340057 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:34Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:34 crc kubenswrapper[4788]: E1010 14:45:34.340235 4788 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.342379 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.342412 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.342426 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.342446 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.342458 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.445068 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.445401 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.445466 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.445554 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.445621 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.547798 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.547846 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.547858 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.547879 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.547891 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.650555 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.650606 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.650619 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.650636 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.650648 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.753122 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.753171 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.753183 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.753199 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.753210 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.861346 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.861383 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.861396 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.861412 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.861422 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.963880 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.964167 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.964270 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.964371 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:34 crc kubenswrapper[4788]: I1010 14:45:34.964485 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:34Z","lastTransitionTime":"2025-10-10T14:45:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.066543 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.066584 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.066594 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.066608 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.066618 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.169314 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.169363 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.169378 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.169396 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.169457 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.233007 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:35 crc kubenswrapper[4788]: E1010 14:45:35.233202 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.272430 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.272472 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.272488 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.272508 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.272523 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.375934 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.375997 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.376007 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.376025 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.376036 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.479204 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.479244 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.479253 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.479268 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.479279 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.582446 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.582860 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.582945 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.583045 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.583160 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.594353 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:35 crc kubenswrapper[4788]: E1010 14:45:35.594548 4788 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:35 crc kubenswrapper[4788]: E1010 14:45:35.594738 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs podName:3d61e9aa-0688-436b-a8af-d6d8353c4350 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:39.594619562 +0000 UTC m=+42.044335140 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs") pod "network-metrics-daemon-jqlwd" (UID: "3d61e9aa-0688-436b-a8af-d6d8353c4350") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.686933 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.687002 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.687017 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.687040 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.687053 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.789686 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.789744 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.789756 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.789774 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.789787 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.893352 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.893412 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.893428 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.893457 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.893475 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.996981 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.997046 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.997062 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.997086 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:35 crc kubenswrapper[4788]: I1010 14:45:35.997099 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:35Z","lastTransitionTime":"2025-10-10T14:45:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.100916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.100974 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.100984 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.101000 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.101010 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.203994 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.204036 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.204044 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.204059 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.204069 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.233582 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.233715 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:36 crc kubenswrapper[4788]: E1010 14:45:36.233733 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:36 crc kubenswrapper[4788]: E1010 14:45:36.233777 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.233788 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:36 crc kubenswrapper[4788]: E1010 14:45:36.233925 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.307743 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.307814 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.307829 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.307852 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.307868 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.410051 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.410109 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.410118 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.410131 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.410159 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.512550 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.512588 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.512600 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.512616 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.512629 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.614455 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.614487 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.614496 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.614510 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.614520 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.716926 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.716959 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.716971 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.716985 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.716995 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.820061 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.820234 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.820258 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.820289 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.820314 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.924160 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.924194 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.924208 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.924227 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:36 crc kubenswrapper[4788]: I1010 14:45:36.924237 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:36Z","lastTransitionTime":"2025-10-10T14:45:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.027513 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.027574 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.027590 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.027608 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.027620 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.131153 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.131205 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.131217 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.131233 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.131243 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.233101 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.233373 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.233407 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.233418 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.233436 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.233447 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: E1010 14:45:37.233517 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.337302 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.337405 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.337432 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.337466 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.337494 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.440236 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.440323 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.440346 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.440376 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.440396 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.543889 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.543974 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.544012 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.544056 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.544082 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.647468 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.647535 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.647549 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.647578 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.647596 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.751082 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.751222 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.751258 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.751273 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.751283 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.855482 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.855546 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.855561 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.855583 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.855596 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.958497 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.958548 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.958561 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.958577 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:37 crc kubenswrapper[4788]: I1010 14:45:37.958589 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:37Z","lastTransitionTime":"2025-10-10T14:45:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.061735 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.061780 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.061788 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.061804 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.061815 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.164731 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.164787 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.164801 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.164823 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.164840 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.233069 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.233184 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.233077 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:38 crc kubenswrapper[4788]: E1010 14:45:38.233409 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:38 crc kubenswrapper[4788]: E1010 14:45:38.233536 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:38 crc kubenswrapper[4788]: E1010 14:45:38.233711 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.264109 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.268175 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.268229 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.268243 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.268264 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.268282 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.281445 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.295058 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.309651 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.331571 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.351184 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.367917 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.372233 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.372275 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.372285 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.372305 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.372318 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.385415 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.405403 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.421057 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.440098 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.455699 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.476424 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.476489 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.476506 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.476528 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.476546 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.476425 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.494237 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.508207 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.523607 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.542559 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:38Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.579764 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.579802 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.579810 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.579831 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.579842 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.682652 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.682695 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.682707 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.682724 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.682736 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.786174 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.786253 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.786268 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.786293 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.786309 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.888134 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.888208 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.888218 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.888234 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.888245 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.991642 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.991689 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.991699 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.991718 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:38 crc kubenswrapper[4788]: I1010 14:45:38.991727 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:38Z","lastTransitionTime":"2025-10-10T14:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.095238 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.095286 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.095296 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.095313 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.095324 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.198653 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.199001 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.199222 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.199374 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.199498 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.233830 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:39 crc kubenswrapper[4788]: E1010 14:45:39.234296 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.302916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.302960 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.302977 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.302997 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.303010 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.406205 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.406321 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.406384 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.406418 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.406474 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.510315 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.510424 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.510456 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.510494 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.510521 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.613924 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.613990 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.614002 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.614023 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.614037 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.640010 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:39 crc kubenswrapper[4788]: E1010 14:45:39.640258 4788 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:39 crc kubenswrapper[4788]: E1010 14:45:39.640360 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs podName:3d61e9aa-0688-436b-a8af-d6d8353c4350 nodeName:}" failed. No retries permitted until 2025-10-10 14:45:47.640333737 +0000 UTC m=+50.090049305 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs") pod "network-metrics-daemon-jqlwd" (UID: "3d61e9aa-0688-436b-a8af-d6d8353c4350") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.716974 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.717039 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.717054 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.717077 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.717094 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.819916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.819964 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.819975 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.819994 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.820005 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.924360 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.924432 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.924451 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.924483 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:39 crc kubenswrapper[4788]: I1010 14:45:39.924503 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:39Z","lastTransitionTime":"2025-10-10T14:45:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.027860 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.027953 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.027973 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.028006 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.028025 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.131035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.131107 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.131120 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.131153 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.131164 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.232880 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.232881 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:40 crc kubenswrapper[4788]: E1010 14:45:40.233200 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:40 crc kubenswrapper[4788]: E1010 14:45:40.233548 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.234104 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.234171 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.234188 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.234208 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.234229 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.233640 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:40 crc kubenswrapper[4788]: E1010 14:45:40.234517 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.338208 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.338648 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.338745 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.338873 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.339045 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.442376 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.442456 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.442493 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.442535 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.442562 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.547326 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.547448 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.547480 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.547523 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.547560 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.650427 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.650458 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.650466 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.650481 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.650489 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.754372 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.754446 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.754506 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.754536 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.754556 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.857750 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.857817 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.857834 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.857859 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.857876 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.960705 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.960738 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.960745 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.960759 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:40 crc kubenswrapper[4788]: I1010 14:45:40.960769 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:40Z","lastTransitionTime":"2025-10-10T14:45:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.064851 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.064931 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.064960 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.065003 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.065033 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.169566 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.169642 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.169662 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.169692 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.169712 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.232984 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:41 crc kubenswrapper[4788]: E1010 14:45:41.233277 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.272627 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.272690 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.272714 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.272746 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.272766 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.377643 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.377723 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.377746 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.377781 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.377806 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.481396 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.481442 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.481457 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.481480 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.481497 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.585366 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.585430 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.585442 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.585461 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.585476 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.688820 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.688855 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.688864 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.688879 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.688888 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.791934 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.792003 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.792015 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.792033 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.792045 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.894580 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.895152 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.895220 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.895293 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.895376 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.998865 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.998916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.998930 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.998949 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:41 crc kubenswrapper[4788]: I1010 14:45:41.998961 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:41Z","lastTransitionTime":"2025-10-10T14:45:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.102025 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.102091 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.102107 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.102133 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.102179 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.205408 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.205469 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.205482 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.205504 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.205519 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.233476 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:42 crc kubenswrapper[4788]: E1010 14:45:42.233626 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.233673 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.233502 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:42 crc kubenswrapper[4788]: E1010 14:45:42.233821 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:42 crc kubenswrapper[4788]: E1010 14:45:42.234218 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.234679 4788 scope.go:117] "RemoveContainer" containerID="5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.308092 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.308482 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.308492 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.308507 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.308516 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.410485 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.410510 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.410517 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.410529 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.410538 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.513155 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.513192 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.513202 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.513216 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.513224 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.604675 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/1.log" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.608331 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.609221 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.615207 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.615373 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.615393 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.615413 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.615429 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.627059 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.649727 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.667182 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.682968 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.697295 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.709210 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.717509 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.717549 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.717557 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.717572 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.717581 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.730635 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.744517 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.761130 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.794495 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.818924 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.820411 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.820450 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.820460 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.820476 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.820486 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.837278 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.850326 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.864950 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.877450 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.888822 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.903923 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:42Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.922553 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.922606 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.922618 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.922637 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:42 crc kubenswrapper[4788]: I1010 14:45:42.922649 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:42Z","lastTransitionTime":"2025-10-10T14:45:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.025533 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.025574 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.025585 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.025603 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.025612 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.127769 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.127813 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.127821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.127835 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.127844 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.230862 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.231124 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.231224 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.231294 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.231367 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.233251 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:43 crc kubenswrapper[4788]: E1010 14:45:43.233381 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.333869 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.333911 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.333920 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.333937 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.333947 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.436582 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.436707 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.436760 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.436789 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.436844 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.539943 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.540005 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.540018 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.540037 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.540050 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.614660 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/2.log" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.615620 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/1.log" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.620041 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04" exitCode=1 Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.620110 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.620177 4788 scope.go:117] "RemoveContainer" containerID="5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.620729 4788 scope.go:117] "RemoveContainer" containerID="062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04" Oct 10 14:45:43 crc kubenswrapper[4788]: E1010 14:45:43.620884 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.636852 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.643752 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.643830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.643850 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.643920 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.643946 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.652766 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.667057 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.685258 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.698670 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.714563 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.730640 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.748024 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.748496 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.748627 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.748749 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.748864 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.758499 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aaf2eaffcb0768ed2f4db132537234287ed17e892ac8c4c682da295844211c0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:29Z\\\",\\\"message\\\":\\\"d already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:29Z is after 2025-08-24T17:21:41Z]\\\\nI1010 14:45:29.587501 6221 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"62af83f3-e0c8-4632-aaaa-17488566a9d8\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/mach\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.772291 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.787026 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.804094 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.817396 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.830540 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.844322 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.851645 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.851679 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.851688 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.851707 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.851717 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.867271 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.881685 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.892998 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:43Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.954098 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.954166 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.954179 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.954200 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:43 crc kubenswrapper[4788]: I1010 14:45:43.954210 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:43Z","lastTransitionTime":"2025-10-10T14:45:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.057040 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.057105 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.057120 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.057165 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.057178 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.159929 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.159974 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.159984 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.159999 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.160009 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.233483 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.233547 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.233631 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.233602 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.233804 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.233920 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.262273 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.262306 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.262315 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.262333 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.262343 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.364817 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.364853 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.364866 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.364882 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.364891 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.468315 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.468364 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.468375 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.468393 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.468407 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.571719 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.571786 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.571808 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.571838 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.571859 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.625426 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/2.log" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.630073 4788 scope.go:117] "RemoveContainer" containerID="062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.630313 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.661175 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.674299 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.674381 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.674404 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.674440 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.674465 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.680788 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.691878 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.707592 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.707641 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.707659 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.707683 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.707698 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.709487 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.723837 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.727878 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.727944 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.727965 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.727990 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.728009 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.731501 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.744423 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.748208 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.757520 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.757558 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.757571 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.757590 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.757603 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.762198 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.770649 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.773564 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.774054 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.774172 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.774265 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.774356 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.774422 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.787592 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.791391 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.791420 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.791429 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.791441 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.791450 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.794183 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.803603 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: E1010 14:45:44.804035 4788 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.805658 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.805707 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.805718 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.805733 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.805759 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.811078 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.838272 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.851889 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.863376 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.874990 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.886239 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.897122 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.908852 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.908856 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:44Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.908890 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.909108 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.909127 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:44 crc kubenswrapper[4788]: I1010 14:45:44.909154 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:44Z","lastTransitionTime":"2025-10-10T14:45:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.011716 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.011760 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.011769 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.011784 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.011796 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.115642 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.115735 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.115752 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.115776 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.115792 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.218780 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.218856 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.218975 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.219069 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.219100 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.233222 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:45 crc kubenswrapper[4788]: E1010 14:45:45.233442 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.322255 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.322309 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.322324 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.322343 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.322362 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.425638 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.425708 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.425725 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.425753 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.425770 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.528524 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.529035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.529218 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.529391 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.529530 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.631715 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.631779 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.631789 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.631804 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.631813 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.733814 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.733861 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.733873 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.733891 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.733902 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.835805 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.835849 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.835862 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.835881 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.835893 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.938115 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.938177 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.938189 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.938204 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:45 crc kubenswrapper[4788]: I1010 14:45:45.938214 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:45Z","lastTransitionTime":"2025-10-10T14:45:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.041180 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.041260 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.041295 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.041325 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.041345 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.144114 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.144199 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.144214 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.144236 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.144250 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.233513 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.233573 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.233513 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:46 crc kubenswrapper[4788]: E1010 14:45:46.233688 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:46 crc kubenswrapper[4788]: E1010 14:45:46.233860 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:46 crc kubenswrapper[4788]: E1010 14:45:46.233978 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.246362 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.246849 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.246996 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.247174 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.247323 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.349549 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.349588 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.349597 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.349615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.349626 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.452306 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.452341 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.452353 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.452367 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.452378 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.554915 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.554952 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.554960 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.554977 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.554985 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.657485 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.657562 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.657585 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.657617 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.657637 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.759715 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.759960 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.760118 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.760224 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.760295 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.862285 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.862322 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.862334 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.862360 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.862373 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.964766 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.964808 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.964819 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.964837 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:46 crc kubenswrapper[4788]: I1010 14:45:46.964848 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:46Z","lastTransitionTime":"2025-10-10T14:45:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.066959 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.067013 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.067022 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.067036 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.067048 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.169065 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.169217 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.169241 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.169267 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.169284 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.233218 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:47 crc kubenswrapper[4788]: E1010 14:45:47.233339 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.271821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.271861 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.271872 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.271890 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.271901 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.373704 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.373752 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.373763 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.373779 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.373789 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.476800 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.476872 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.476885 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.476905 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.476915 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.579679 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.579741 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.579753 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.579774 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.579811 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.685606 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.685678 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.685690 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.685726 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.685736 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.730799 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:47 crc kubenswrapper[4788]: E1010 14:45:47.730991 4788 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:47 crc kubenswrapper[4788]: E1010 14:45:47.731046 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs podName:3d61e9aa-0688-436b-a8af-d6d8353c4350 nodeName:}" failed. No retries permitted until 2025-10-10 14:46:03.731031773 +0000 UTC m=+66.180747311 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs") pod "network-metrics-daemon-jqlwd" (UID: "3d61e9aa-0688-436b-a8af-d6d8353c4350") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.788916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.788953 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.788963 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.788992 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.789001 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.892349 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.892408 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.892417 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.892431 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.892444 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.996635 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.996719 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.996737 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.996760 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:47 crc kubenswrapper[4788]: I1010 14:45:47.996776 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:47Z","lastTransitionTime":"2025-10-10T14:45:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.099828 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.099865 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.099874 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.099888 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.099899 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.202193 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.202263 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.202281 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.202303 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.202323 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.233710 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.233823 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.233925 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:48 crc kubenswrapper[4788]: E1010 14:45:48.233923 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:48 crc kubenswrapper[4788]: E1010 14:45:48.234023 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:48 crc kubenswrapper[4788]: E1010 14:45:48.234211 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.251509 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.264028 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.289551 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.299110 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.304442 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.304484 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.304509 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.304525 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.304534 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.310694 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.327169 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.340106 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.352222 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.368569 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.421973 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.423325 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.423373 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.423388 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.423406 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.423415 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.440725 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.453991 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.466436 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.478380 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.493929 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.514767 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.518934 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.528476 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.528515 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.528527 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.528545 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.528558 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.528588 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.533510 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.544309 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.556591 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.573017 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.583409 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.595593 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.608302 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.618400 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.630404 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.630436 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.630445 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.630460 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.630469 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.637967 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.647032 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.659748 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.670076 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.681647 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.692889 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.705810 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.728528 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.732423 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.732453 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.732464 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.732478 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.732488 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.742939 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.755311 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:48Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.835229 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.835267 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.835278 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.835298 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.835310 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.938405 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.938645 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.938771 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.938865 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:48 crc kubenswrapper[4788]: I1010 14:45:48.938948 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:48Z","lastTransitionTime":"2025-10-10T14:45:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.041389 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.041445 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.041461 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.041481 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.041496 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.143680 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.143733 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.143756 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.143784 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.143805 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.233233 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.233397 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.246102 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.246207 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.246227 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.246250 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.246266 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.349072 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.349119 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.349131 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.349165 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.349175 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.451686 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.451956 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.452113 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.452280 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.452412 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.555403 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.555730 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.555758 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.555776 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.555789 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.658363 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.658404 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.658421 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.658440 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.658451 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.760427 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.760475 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.760488 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.760505 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.760517 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.862777 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.862817 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.862827 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.862842 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.862852 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.953500 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.953654 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.953781 4788 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.953804 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:46:21.953742479 +0000 UTC m=+84.403458047 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.953891 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:46:21.953872023 +0000 UTC m=+84.403587581 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.953928 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.953974 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.954086 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.954104 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.954117 4788 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.954173 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 14:46:21.9541634 +0000 UTC m=+84.403878968 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.954282 4788 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:49 crc kubenswrapper[4788]: E1010 14:45:49.954331 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:46:21.954319654 +0000 UTC m=+84.404035212 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.964889 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.964916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.964926 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.964944 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:49 crc kubenswrapper[4788]: I1010 14:45:49.964954 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:49Z","lastTransitionTime":"2025-10-10T14:45:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.055422 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:50 crc kubenswrapper[4788]: E1010 14:45:50.055606 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:45:50 crc kubenswrapper[4788]: E1010 14:45:50.055630 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:45:50 crc kubenswrapper[4788]: E1010 14:45:50.055644 4788 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:50 crc kubenswrapper[4788]: E1010 14:45:50.055696 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 14:46:22.055677611 +0000 UTC m=+84.505393179 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.066896 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.067116 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.067222 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.067313 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.067393 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.169686 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.169755 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.169773 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.169792 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.169807 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.233466 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.233533 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.233471 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:50 crc kubenswrapper[4788]: E1010 14:45:50.233598 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:50 crc kubenswrapper[4788]: E1010 14:45:50.233802 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:50 crc kubenswrapper[4788]: E1010 14:45:50.233894 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.272198 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.272277 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.272295 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.272319 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.272337 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.375986 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.376065 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.376076 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.376103 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.376116 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.478296 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.478344 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.478362 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.478384 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.478399 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.580546 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.580590 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.580607 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.580624 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.580638 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.683191 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.683230 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.683243 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.683260 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.683273 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.785728 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.785777 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.785789 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.785806 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.785817 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.887598 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.887651 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.887660 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.887673 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.887682 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.989677 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.989712 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.989722 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.989737 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:50 crc kubenswrapper[4788]: I1010 14:45:50.989746 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:50Z","lastTransitionTime":"2025-10-10T14:45:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.092481 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.092512 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.092520 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.092532 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.092543 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.195560 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.195607 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.195624 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.195683 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.195701 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.233811 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:51 crc kubenswrapper[4788]: E1010 14:45:51.233976 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.298481 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.298535 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.298546 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.298565 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.298577 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.401673 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.401709 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.401717 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.401730 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.401739 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.503852 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.503937 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.503952 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.503967 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.503978 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.607186 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.607246 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.607263 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.607284 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.607301 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.709524 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.709561 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.709571 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.709587 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.709598 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.811626 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.811682 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.811699 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.811722 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.811736 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.913742 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.913780 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.913792 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.913812 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:51 crc kubenswrapper[4788]: I1010 14:45:51.913825 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:51Z","lastTransitionTime":"2025-10-10T14:45:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.016456 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.016507 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.016529 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.016553 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.016568 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.119119 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.119178 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.119192 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.119228 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.119240 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.221416 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.221454 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.221463 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.221478 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.221488 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.233660 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.233712 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:52 crc kubenswrapper[4788]: E1010 14:45:52.233771 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.233678 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:52 crc kubenswrapper[4788]: E1010 14:45:52.233899 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:52 crc kubenswrapper[4788]: E1010 14:45:52.233958 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.324178 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.324219 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.324228 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.324242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.324252 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.426465 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.426493 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.426501 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.426513 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.426524 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.529090 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.529124 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.529134 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.529170 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.529181 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.632116 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.632170 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.632182 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.632198 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.632208 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.734659 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.734693 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.734725 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.734740 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.734750 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.837323 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.837348 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.837355 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.837367 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.837375 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.940131 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.940238 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.940255 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.940277 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:52 crc kubenswrapper[4788]: I1010 14:45:52.940294 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:52Z","lastTransitionTime":"2025-10-10T14:45:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.042549 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.042643 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.042677 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.042708 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.042731 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.144945 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.145026 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.145060 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.145093 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.145115 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.233181 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:53 crc kubenswrapper[4788]: E1010 14:45:53.233371 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.247182 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.247203 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.247210 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.247222 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.247231 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.349587 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.349646 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.349663 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.349686 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.349703 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.451884 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.451928 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.451937 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.451953 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.451962 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.554583 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.554618 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.554627 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.554640 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.554648 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.657545 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.657598 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.657614 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.657635 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.657650 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.759700 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.759751 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.759768 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.759788 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.759806 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.861921 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.861959 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.861970 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.861986 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.861997 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.964699 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.964792 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.964805 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.964821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:53 crc kubenswrapper[4788]: I1010 14:45:53.964832 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:53Z","lastTransitionTime":"2025-10-10T14:45:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.067611 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.067650 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.067662 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.067699 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.067709 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.170777 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.170826 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.170843 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.170865 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.170879 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.233560 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.233562 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.233627 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:54 crc kubenswrapper[4788]: E1010 14:45:54.233743 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:54 crc kubenswrapper[4788]: E1010 14:45:54.233839 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:54 crc kubenswrapper[4788]: E1010 14:45:54.234245 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.272912 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.272989 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.273003 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.273022 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.273035 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.375658 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.375872 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.375915 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.375950 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.375973 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.478976 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.479104 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.479125 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.479173 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.479202 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.581332 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.581383 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.581394 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.581413 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.581424 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.685050 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.685082 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.685093 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.685109 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.685119 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.788413 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.788478 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.788493 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.788510 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.788522 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.891975 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.892044 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.892067 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.892100 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.892122 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.994947 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.995018 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.995042 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.995067 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:54 crc kubenswrapper[4788]: I1010 14:45:54.995085 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:54Z","lastTransitionTime":"2025-10-10T14:45:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.098242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.098280 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.098292 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.098306 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.098317 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.152327 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.152414 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.152433 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.152466 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.152496 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: E1010 14:45:55.175181 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:55Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.180697 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.180771 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.180786 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.180815 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.180835 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: E1010 14:45:55.201155 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:55Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.206917 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.206960 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.206977 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.206998 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.207011 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: E1010 14:45:55.220121 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:55Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.224693 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.224739 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.224755 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.224774 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.224790 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.233244 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:55 crc kubenswrapper[4788]: E1010 14:45:55.233399 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:55 crc kubenswrapper[4788]: E1010 14:45:55.240986 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:55Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.244405 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.244447 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.244466 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.244487 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.244502 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: E1010 14:45:55.257057 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:55Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:55 crc kubenswrapper[4788]: E1010 14:45:55.257261 4788 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.258776 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.258818 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.258841 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.258863 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.258879 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.361103 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.361195 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.361215 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.361233 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.361263 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.463821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.463871 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.463886 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.463906 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.463921 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.566653 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.566695 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.566705 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.566723 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.566734 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.669363 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.669436 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.669456 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.669478 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.669496 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.772696 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.772757 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.772777 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.772803 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.772821 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.876476 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.877202 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.877269 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.877301 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.877313 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.980347 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.980406 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.980417 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.980434 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:55 crc kubenswrapper[4788]: I1010 14:45:55.980444 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:55Z","lastTransitionTime":"2025-10-10T14:45:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.083515 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.083565 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.083582 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.083607 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.083626 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.187324 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.187383 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.187393 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.187408 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.187417 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.233686 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:56 crc kubenswrapper[4788]: E1010 14:45:56.233872 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.234182 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:56 crc kubenswrapper[4788]: E1010 14:45:56.234276 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.234296 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:56 crc kubenswrapper[4788]: E1010 14:45:56.234385 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.235014 4788 scope.go:117] "RemoveContainer" containerID="062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04" Oct 10 14:45:56 crc kubenswrapper[4788]: E1010 14:45:56.235173 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.290118 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.290229 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.290272 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.290304 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.290322 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.393435 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.393506 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.393525 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.393551 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.393567 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.496513 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.496557 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.496567 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.496582 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.496594 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.599599 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.599642 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.599653 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.599667 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.599677 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.701460 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.701509 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.701521 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.701539 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.701549 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.804264 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.804315 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.804333 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.804354 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.804371 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.906777 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.906861 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.906886 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.906915 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:56 crc kubenswrapper[4788]: I1010 14:45:56.906931 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:56Z","lastTransitionTime":"2025-10-10T14:45:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.009324 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.009394 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.009411 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.009432 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.009447 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.111662 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.111725 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.111742 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.111764 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.111783 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.214065 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.214111 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.214124 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.214156 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.214171 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.233552 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:57 crc kubenswrapper[4788]: E1010 14:45:57.233714 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.316396 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.316442 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.316451 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.316467 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.316477 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.418482 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.418515 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.418523 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.418538 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.418549 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.520361 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.520406 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.520420 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.520436 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.520446 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.622918 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.622968 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.622978 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.622992 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.623002 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.724831 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.724866 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.724878 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.724894 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.724904 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.827624 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.827657 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.827668 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.827685 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.827697 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.930906 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.930942 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.930950 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.930963 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:57 crc kubenswrapper[4788]: I1010 14:45:57.930972 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:57Z","lastTransitionTime":"2025-10-10T14:45:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.033686 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.033751 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.033762 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.033779 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.033795 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.135891 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.135930 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.135941 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.135955 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.135964 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.232888 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.232888 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.233052 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:45:58 crc kubenswrapper[4788]: E1010 14:45:58.233090 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:45:58 crc kubenswrapper[4788]: E1010 14:45:58.232988 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:45:58 crc kubenswrapper[4788]: E1010 14:45:58.233241 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.237875 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.237926 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.237943 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.237962 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.237978 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.247045 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.263961 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.274614 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.286211 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.299278 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.312957 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.331484 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.339564 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.339587 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.339596 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.339609 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.339619 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.351384 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.362110 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.378979 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.392256 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.404550 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.415304 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.425282 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.435673 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.441013 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.441043 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.441051 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.441064 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.441073 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.452278 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.462665 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.471592 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:45:58Z is after 2025-08-24T17:21:41Z" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.542561 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.542598 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.542607 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.542620 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.542629 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.645005 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.645039 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.645050 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.645064 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.645073 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.747613 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.747649 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.747657 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.747674 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.747682 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.850069 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.850121 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.850132 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.850168 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.850180 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.952875 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.952920 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.952936 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.952956 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:58 crc kubenswrapper[4788]: I1010 14:45:58.952972 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:58Z","lastTransitionTime":"2025-10-10T14:45:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.055740 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.055778 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.055802 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.055815 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.055825 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.158775 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.158814 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.158826 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.158839 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.158849 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.233031 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:45:59 crc kubenswrapper[4788]: E1010 14:45:59.233283 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.261001 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.261039 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.261049 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.261063 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.261075 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.363518 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.363585 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.363600 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.363624 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.363641 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.466687 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.466733 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.466742 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.466755 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.466763 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.569175 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.569232 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.569241 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.569255 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.569264 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.671669 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.671704 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.671714 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.671729 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.671740 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.775889 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.775940 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.775948 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.775966 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.775975 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.879023 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.879086 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.879102 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.879126 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.879165 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.982519 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.982571 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.982587 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.982613 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:45:59 crc kubenswrapper[4788]: I1010 14:45:59.982630 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:45:59Z","lastTransitionTime":"2025-10-10T14:45:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.085753 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.085811 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.085829 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.085852 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.085869 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.190006 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.190071 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.190091 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.190115 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.190134 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.233004 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.233019 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:00 crc kubenswrapper[4788]: E1010 14:46:00.233187 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.233250 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:00 crc kubenswrapper[4788]: E1010 14:46:00.233323 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:00 crc kubenswrapper[4788]: E1010 14:46:00.233478 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.293749 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.293842 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.293875 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.293909 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.293934 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.396209 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.396248 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.396262 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.396280 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.396293 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.498435 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.498474 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.498485 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.498498 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.498507 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.601002 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.601059 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.601072 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.601097 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.601110 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.703943 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.703981 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.703992 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.704020 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.704032 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.806783 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.806824 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.806839 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.806860 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.806875 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.909293 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.909327 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.909336 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.909349 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:00 crc kubenswrapper[4788]: I1010 14:46:00.909360 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:00Z","lastTransitionTime":"2025-10-10T14:46:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.011283 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.011313 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.011320 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.011333 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.011341 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.113819 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.113869 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.113880 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.113897 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.113909 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.217130 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.217212 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.217225 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.217248 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.217264 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.233788 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:01 crc kubenswrapper[4788]: E1010 14:46:01.233938 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.319895 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.319942 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.319955 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.319975 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.319986 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.422784 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.422881 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.422894 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.422918 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.422936 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.526872 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.526940 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.526963 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.526997 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.527017 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.631377 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.631439 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.631459 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.631489 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.631512 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.735033 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.735104 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.735121 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.735180 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.735200 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.840044 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.840130 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.840191 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.840227 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.840249 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.943522 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.943588 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.943608 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.943634 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:01 crc kubenswrapper[4788]: I1010 14:46:01.943657 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:01Z","lastTransitionTime":"2025-10-10T14:46:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.045900 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.045949 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.045959 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.045973 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.045982 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.147977 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.148011 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.148021 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.148036 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.148045 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.232861 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:02 crc kubenswrapper[4788]: E1010 14:46:02.233007 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.233031 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.232892 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:02 crc kubenswrapper[4788]: E1010 14:46:02.233279 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:02 crc kubenswrapper[4788]: E1010 14:46:02.233361 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.250571 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.250601 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.250613 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.250630 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.250643 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.354743 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.354807 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.354821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.354842 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.354856 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.458201 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.459160 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.459319 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.459487 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.459635 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.562173 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.562222 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.562231 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.562265 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.562275 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.665256 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.665313 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.665323 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.665344 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.665359 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.768805 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.768864 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.768875 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.768891 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.768902 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.873309 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.873877 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.874079 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.874365 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.874623 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.978707 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.978776 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.978796 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.978877 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:02 crc kubenswrapper[4788]: I1010 14:46:02.978897 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:02Z","lastTransitionTime":"2025-10-10T14:46:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.081738 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.081781 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.081792 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.081812 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.081826 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.184615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.184703 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.184732 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.184764 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.184788 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.233363 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:03 crc kubenswrapper[4788]: E1010 14:46:03.233513 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.287465 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.287515 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.287525 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.287539 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.287548 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.389544 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.389599 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.389612 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.389639 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.389653 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.492083 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.492155 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.492169 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.492185 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.492198 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.595245 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.595319 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.595331 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.595354 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.595364 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.699712 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.699772 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.699785 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.699809 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.699827 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.788637 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:03 crc kubenswrapper[4788]: E1010 14:46:03.788858 4788 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:46:03 crc kubenswrapper[4788]: E1010 14:46:03.788960 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs podName:3d61e9aa-0688-436b-a8af-d6d8353c4350 nodeName:}" failed. No retries permitted until 2025-10-10 14:46:35.788933689 +0000 UTC m=+98.238649247 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs") pod "network-metrics-daemon-jqlwd" (UID: "3d61e9aa-0688-436b-a8af-d6d8353c4350") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.802683 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.802752 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.802768 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.802798 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.802822 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.906259 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.906310 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.906321 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.906338 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:03 crc kubenswrapper[4788]: I1010 14:46:03.906349 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:03Z","lastTransitionTime":"2025-10-10T14:46:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.009337 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.009830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.009839 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.009858 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.009873 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.112446 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.112511 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.112522 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.112543 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.112555 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.216177 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.216230 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.216244 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.216267 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.216285 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.233691 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:04 crc kubenswrapper[4788]: E1010 14:46:04.233903 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.233691 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.233958 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:04 crc kubenswrapper[4788]: E1010 14:46:04.234074 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:04 crc kubenswrapper[4788]: E1010 14:46:04.234288 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.319772 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.319860 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.319873 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.319901 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.319917 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.423625 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.424316 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.424370 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.424399 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.424418 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.527391 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.527459 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.527485 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.527512 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.527535 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.630360 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.630421 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.630439 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.630464 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.630478 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.732991 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.733036 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.733047 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.733062 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.733070 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.835821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.835861 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.835869 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.835886 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.835897 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.939190 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.939262 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.939281 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.939309 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:04 crc kubenswrapper[4788]: I1010 14:46:04.939328 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:04Z","lastTransitionTime":"2025-10-10T14:46:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.041646 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.041696 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.041709 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.041726 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.041796 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.144520 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.144565 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.144574 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.144592 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.144602 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.232974 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:05 crc kubenswrapper[4788]: E1010 14:46:05.233237 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.247978 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.248016 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.248026 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.248043 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.248055 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.351100 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.351229 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.351247 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.351269 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.351285 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.360663 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.360732 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.360753 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.360783 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.360805 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: E1010 14:46:05.387182 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:05Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.393112 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.393235 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.393260 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.393293 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.393317 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: E1010 14:46:05.414425 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:05Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.420509 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.420580 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.420603 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.420635 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.420654 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: E1010 14:46:05.441761 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:05Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.447348 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.447756 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.447847 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.447958 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.448048 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: E1010 14:46:05.462293 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:05Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.467351 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.467433 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.467454 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.467484 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.467500 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: E1010 14:46:05.484053 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:05Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:05 crc kubenswrapper[4788]: E1010 14:46:05.484242 4788 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.486067 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.486108 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.486122 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.486164 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.486181 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.589870 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.589937 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.589952 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.589975 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.589989 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.692718 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.692796 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.692810 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.692830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.692846 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.796232 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.796277 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.796287 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.796304 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.796313 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.899710 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.899756 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.899767 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.899785 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:05 crc kubenswrapper[4788]: I1010 14:46:05.899796 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:05Z","lastTransitionTime":"2025-10-10T14:46:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.002792 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.002863 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.002876 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.002899 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.002913 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.105940 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.106487 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.106664 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.106826 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.106971 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.209810 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.209891 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.209915 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.209953 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.209980 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.233993 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.233993 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.234383 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:06 crc kubenswrapper[4788]: E1010 14:46:06.234485 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:06 crc kubenswrapper[4788]: E1010 14:46:06.234603 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:06 crc kubenswrapper[4788]: E1010 14:46:06.234755 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.312708 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.312772 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.312782 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.312796 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.312806 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.414858 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.414895 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.414908 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.414925 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.414937 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.518354 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.518433 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.518452 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.518487 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.518512 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.621096 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.621170 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.621184 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.621202 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.621217 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.705342 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/0.log" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.705398 4788 generic.go:334] "Generic (PLEG): container finished" podID="39cfa1ec-4912-4550-9ead-7f6113db2221" containerID="7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5" exitCode=1 Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.705425 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rzgfk" event={"ID":"39cfa1ec-4912-4550-9ead-7f6113db2221","Type":"ContainerDied","Data":"7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.705783 4788 scope.go:117] "RemoveContainer" containerID="7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.719873 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.734855 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.734917 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.734935 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.734956 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.734975 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.755173 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.773611 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.791889 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.809050 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.827379 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.840187 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.840474 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.840605 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.840693 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.840774 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.849264 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.865108 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"2025-10-10T14:45:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44\\\\n2025-10-10T14:45:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44 to /host/opt/cni/bin/\\\\n2025-10-10T14:45:20Z [verbose] multus-daemon started\\\\n2025-10-10T14:45:20Z [verbose] Readiness Indicator file check\\\\n2025-10-10T14:46:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.881964 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.897181 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.914749 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.926191 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.943720 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.943953 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.944031 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.944100 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.944214 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:06Z","lastTransitionTime":"2025-10-10T14:46:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.951086 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.964715 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.978860 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:06 crc kubenswrapper[4788]: I1010 14:46:06.991049 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:06Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.003182 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.013971 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.047495 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.047550 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.047570 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.047597 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.047612 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.150741 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.150783 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.150794 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.150812 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.150824 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.233701 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:07 crc kubenswrapper[4788]: E1010 14:46:07.233856 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.254894 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.254954 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.254969 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.254990 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.255004 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.357591 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.357692 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.357715 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.357747 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.357772 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.460004 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.460040 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.460053 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.460072 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.460090 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.562643 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.562683 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.562693 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.562706 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.562716 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.665385 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.665422 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.665433 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.665449 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.665459 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.711164 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/0.log" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.711217 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rzgfk" event={"ID":"39cfa1ec-4912-4550-9ead-7f6113db2221","Type":"ContainerStarted","Data":"a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.726624 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.742920 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.755908 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.767830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.767890 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.767904 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.767925 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.767938 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.771029 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.781433 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.794814 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.816839 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.828227 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.847275 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.861545 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.869697 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.869763 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.869780 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.869797 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.869809 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.874319 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.885771 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.897400 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"2025-10-10T14:45:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44\\\\n2025-10-10T14:45:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44 to /host/opt/cni/bin/\\\\n2025-10-10T14:45:20Z [verbose] multus-daemon started\\\\n2025-10-10T14:45:20Z [verbose] Readiness Indicator file check\\\\n2025-10-10T14:46:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:46:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.908104 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.921265 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.939974 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.950706 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.969028 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:07Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.971373 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.971404 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.971414 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.971428 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:07 crc kubenswrapper[4788]: I1010 14:46:07.971437 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:07Z","lastTransitionTime":"2025-10-10T14:46:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.077361 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.078904 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.079015 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.079037 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.079049 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.181541 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.181582 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.181594 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.181609 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.181618 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.233441 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.233524 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:08 crc kubenswrapper[4788]: E1010 14:46:08.233582 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:08 crc kubenswrapper[4788]: E1010 14:46:08.233648 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.233521 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:08 crc kubenswrapper[4788]: E1010 14:46:08.233743 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.285546 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.285598 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.285613 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.285634 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.285647 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.285547 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.310553 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.328862 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.342070 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.354972 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.366639 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.377677 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.388829 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.388975 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.389005 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.389015 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.389030 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.389039 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.404405 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.418494 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.443895 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.453573 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.468488 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.483840 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.492484 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.492531 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.492545 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.492570 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.492586 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.496388 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.515247 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.532339 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"2025-10-10T14:45:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44\\\\n2025-10-10T14:45:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44 to /host/opt/cni/bin/\\\\n2025-10-10T14:45:20Z [verbose] multus-daemon started\\\\n2025-10-10T14:45:20Z [verbose] Readiness Indicator file check\\\\n2025-10-10T14:46:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:46:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.549634 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:08Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.594106 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.594154 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.594166 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.594182 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.594193 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.700001 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.700241 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.700283 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.700305 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.700324 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.803162 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.803209 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.803219 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.803234 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.803245 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.905870 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.905915 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.905929 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.905947 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:08 crc kubenswrapper[4788]: I1010 14:46:08.905962 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:08Z","lastTransitionTime":"2025-10-10T14:46:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.008504 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.008564 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.008584 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.008610 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.008627 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.111601 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.111655 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.111666 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.111681 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.111692 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.214763 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.214837 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.214855 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.214881 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.214898 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.233360 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:09 crc kubenswrapper[4788]: E1010 14:46:09.233523 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.317332 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.317387 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.317398 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.317418 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.317433 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.420953 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.420993 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.421001 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.421014 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.421023 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.523713 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.523748 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.523758 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.523772 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.523782 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.625999 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.626032 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.626041 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.626054 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.626063 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.727900 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.727941 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.727951 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.727964 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.727974 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.829857 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.829889 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.829897 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.829912 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.829922 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.932615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.932655 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.932666 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.932682 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:09 crc kubenswrapper[4788]: I1010 14:46:09.932693 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:09Z","lastTransitionTime":"2025-10-10T14:46:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.035218 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.035252 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.035261 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.035274 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.035283 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.137832 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.137877 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.137887 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.137902 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.137910 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.233459 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.233517 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.233571 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:10 crc kubenswrapper[4788]: E1010 14:46:10.233638 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:10 crc kubenswrapper[4788]: E1010 14:46:10.233723 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:10 crc kubenswrapper[4788]: E1010 14:46:10.233795 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.239550 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.239594 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.239602 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.239617 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.239626 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.342859 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.342932 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.342952 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.342982 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.343001 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.445959 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.446034 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.446056 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.446125 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.446192 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.548392 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.548444 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.548457 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.548475 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.548486 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.651042 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.651076 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.651090 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.651106 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.651114 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.752970 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.753010 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.753022 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.753041 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.753053 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.855424 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.855700 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.855717 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.855732 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.855742 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.957448 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.957706 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.957778 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.957840 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:10 crc kubenswrapper[4788]: I1010 14:46:10.957895 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:10Z","lastTransitionTime":"2025-10-10T14:46:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.059706 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.059950 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.060024 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.060109 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.060237 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.162505 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.162559 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.162572 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.162590 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.162602 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.233429 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.233947 4788 scope.go:117] "RemoveContainer" containerID="062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04" Oct 10 14:46:11 crc kubenswrapper[4788]: E1010 14:46:11.233946 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.265861 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.266203 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.266219 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.266254 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.266266 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.368439 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.368462 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.368486 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.368499 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.368507 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.470658 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.470699 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.470709 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.470724 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.470733 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.572853 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.572890 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.572901 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.572949 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.572962 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.675821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.675860 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.675872 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.675886 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.675897 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.724324 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/2.log" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.726554 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.727126 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.743480 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.753358 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.762095 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.772914 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.777346 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.777383 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.777394 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.777407 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.777416 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.785818 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.796424 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.805491 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.820378 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.832198 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.842239 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.859016 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:46:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.868333 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.879830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.879857 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.879864 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.879878 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.879888 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.879892 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.893214 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.903275 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.915311 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.925744 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"2025-10-10T14:45:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44\\\\n2025-10-10T14:45:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44 to /host/opt/cni/bin/\\\\n2025-10-10T14:45:20Z [verbose] multus-daemon started\\\\n2025-10-10T14:45:20Z [verbose] Readiness Indicator file check\\\\n2025-10-10T14:46:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:46:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.937881 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:11Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.982629 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.982688 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.982704 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.982724 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:11 crc kubenswrapper[4788]: I1010 14:46:11.982737 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:11Z","lastTransitionTime":"2025-10-10T14:46:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.085401 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.085438 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.085448 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.085462 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.085471 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.188023 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.188093 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.188116 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.188178 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.188201 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.233714 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.233744 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.233728 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:12 crc kubenswrapper[4788]: E1010 14:46:12.233920 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:12 crc kubenswrapper[4788]: E1010 14:46:12.234002 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:12 crc kubenswrapper[4788]: E1010 14:46:12.234105 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.291280 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.291314 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.291323 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.291338 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.291350 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.394940 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.394982 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.394990 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.395007 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.395015 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.498299 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.498341 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.498356 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.498376 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.498392 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.600934 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.600978 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.600990 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.601006 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.601017 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.703736 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.703803 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.703821 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.703847 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.703866 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.732159 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/3.log" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.733659 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/2.log" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.738880 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" exitCode=1 Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.738937 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.739020 4788 scope.go:117] "RemoveContainer" containerID="062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.740231 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:46:12 crc kubenswrapper[4788]: E1010 14:46:12.740568 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.767769 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.779606 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.792645 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.805873 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.805909 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.805919 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.805935 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.805946 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.806135 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.823256 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.836035 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.849721 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.859584 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.878799 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.890040 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.908808 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.908863 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.908875 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.908892 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.908905 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:12Z","lastTransitionTime":"2025-10-10T14:46:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.910992 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://062f31f5ccc92d6f4f3a56c166f3624902e70cf0a65b7cd3f254ee7ae7c15d04\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:45:43Z\\\",\\\"message\\\":\\\"-gdk6g openshift-network-diagnostics/network-check-target-xd92c openshift-ovn-kubernetes/ovnkube-node-cwgns openshift-etcd/etcd-crc openshift-kube-apiserver/kube-apiserver-crc openshift-network-operator/network-operator-58b4c7f79c-55gtf]\\\\nI1010 14:45:43.206587 6421 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1010 14:45:43.206601 6421 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206612 6421 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206621 6421 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf in node crc\\\\nI1010 14:45:43.206628 6421 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf after 0 failed attempt(s)\\\\nI1010 14:45:43.206632 6421 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1010 14:45:43.206647 6421 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1010 14:45:43.206699 6421 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:12Z\\\",\\\"message\\\":\\\"nsole]} name:Service_openshift-console/console_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.194:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d7d7b270-1480-47f8-bdf9-690dbab310cb}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1010 14:46:12.051320 6784 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nI1010 14:46:12.051326 6784 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nI1010 14:46:12.051330 6784 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx in node crc\\\\nI1010 14:46:12.051334 6784 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx after 0 failed attempt(s)\\\\nI1010 14:46:12.051338 6784 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nF1010 14:46:12.051352 6784 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:46:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.920990 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.932189 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.943421 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.955330 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.970995 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.984564 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"2025-10-10T14:45:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44\\\\n2025-10-10T14:45:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44 to /host/opt/cni/bin/\\\\n2025-10-10T14:45:20Z [verbose] multus-daemon started\\\\n2025-10-10T14:45:20Z [verbose] Readiness Indicator file check\\\\n2025-10-10T14:46:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:46:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:12 crc kubenswrapper[4788]: I1010 14:46:12.998306 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:12Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.011360 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.011391 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.011399 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.011418 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.011429 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.114387 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.114440 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.114454 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.114476 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.114493 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.217624 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.217659 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.217670 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.217687 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.217699 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.233501 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:13 crc kubenswrapper[4788]: E1010 14:46:13.233666 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.320379 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.320435 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.320444 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.320466 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.320475 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.423247 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.423351 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.423376 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.423857 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.424130 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.527995 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.528047 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.528064 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.528129 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.528188 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.632104 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.632184 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.632222 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.632249 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.632269 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.736427 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.736479 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.736490 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.736506 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.736516 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.745950 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/3.log" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.751568 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:46:13 crc kubenswrapper[4788]: E1010 14:46:13.751773 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.780399 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.801451 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.816764 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.832045 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.839173 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.839230 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.839248 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.839273 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.839289 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.856441 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.870588 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.889430 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.903316 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.918409 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.938483 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.941557 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.941619 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.941631 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.941654 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.941668 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:13Z","lastTransitionTime":"2025-10-10T14:46:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.959065 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:12Z\\\",\\\"message\\\":\\\"nsole]} name:Service_openshift-console/console_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.194:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d7d7b270-1480-47f8-bdf9-690dbab310cb}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1010 14:46:12.051320 6784 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nI1010 14:46:12.051326 6784 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nI1010 14:46:12.051330 6784 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx in node crc\\\\nI1010 14:46:12.051334 6784 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx after 0 failed attempt(s)\\\\nI1010 14:46:12.051338 6784 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nF1010 14:46:12.051352 6784 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:46:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.972029 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:13 crc kubenswrapper[4788]: I1010 14:46:13.986451 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.001997 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:13Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.017641 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:14Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.031372 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:14Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.043815 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.043861 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.043878 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.043900 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.043915 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.051326 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"2025-10-10T14:45:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44\\\\n2025-10-10T14:45:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44 to /host/opt/cni/bin/\\\\n2025-10-10T14:45:20Z [verbose] multus-daemon started\\\\n2025-10-10T14:45:20Z [verbose] Readiness Indicator file check\\\\n2025-10-10T14:46:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:46:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:14Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.064802 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:14Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.146565 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.146615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.146630 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.146646 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.146657 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.233569 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.233629 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:14 crc kubenswrapper[4788]: E1010 14:46:14.233717 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.233797 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:14 crc kubenswrapper[4788]: E1010 14:46:14.233932 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:14 crc kubenswrapper[4788]: E1010 14:46:14.234305 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.249678 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.249715 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.249751 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.249765 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.249777 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.353076 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.353168 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.353187 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.353207 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.353219 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.455888 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.455944 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.455958 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.455978 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.455990 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.558876 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.558943 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.558961 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.558983 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.558998 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.662188 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.662232 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.662242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.662255 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.662265 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.765044 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.765081 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.765093 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.765132 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.765160 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.867859 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.867898 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.867906 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.867919 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.867932 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.970648 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.970682 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.970689 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.970702 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:14 crc kubenswrapper[4788]: I1010 14:46:14.970710 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:14Z","lastTransitionTime":"2025-10-10T14:46:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.073430 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.073482 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.073492 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.073509 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.073521 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.176411 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.176733 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.176866 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.177001 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.177128 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.233478 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:15 crc kubenswrapper[4788]: E1010 14:46:15.233750 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.279396 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.279658 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.279752 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.279865 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.279969 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.383564 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.383720 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.383750 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.383784 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.383806 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.486169 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.486470 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.486539 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.486613 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.486746 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.588873 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.588916 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.588928 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.588944 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.588957 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.692431 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.692499 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.692530 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.692554 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.692572 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.693958 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.694005 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.694033 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.694054 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.694069 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: E1010 14:46:15.707029 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:15Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.711334 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.711406 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.711429 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.711456 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.711479 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: E1010 14:46:15.725870 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:15Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.730076 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.730110 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.730121 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.730154 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.730166 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: E1010 14:46:15.747648 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:15Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.752231 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.752273 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.752285 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.752300 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.752311 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: E1010 14:46:15.765227 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:15Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.769164 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.769205 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.769215 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.769229 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.769239 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: E1010 14:46:15.786832 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:15Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:15 crc kubenswrapper[4788]: E1010 14:46:15.786943 4788 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.794303 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.794525 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.794669 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.794804 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.794979 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.898441 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.898476 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.898487 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.898502 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:15 crc kubenswrapper[4788]: I1010 14:46:15.898516 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:15Z","lastTransitionTime":"2025-10-10T14:46:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.001175 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.001207 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.001217 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.001231 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.001241 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.103894 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.103957 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.103980 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.104008 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.104028 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.206704 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.206747 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.206757 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.206774 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.206785 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.233824 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.233866 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.234047 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:16 crc kubenswrapper[4788]: E1010 14:46:16.234041 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:16 crc kubenswrapper[4788]: E1010 14:46:16.234170 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:16 crc kubenswrapper[4788]: E1010 14:46:16.234242 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.310642 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.310711 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.310738 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.310771 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.310794 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.413694 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.413789 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.413810 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.413836 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.413854 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.516696 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.516773 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.516796 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.516825 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.516843 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.619539 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.619623 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.619635 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.619649 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.619658 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.725306 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.725358 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.725433 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.725988 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.726002 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.828188 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.828224 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.828233 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.828247 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.828257 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.930181 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.930212 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.930224 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.930240 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:16 crc kubenswrapper[4788]: I1010 14:46:16.930251 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:16Z","lastTransitionTime":"2025-10-10T14:46:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.032598 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.032634 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.032644 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.032660 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.032670 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.135779 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.135829 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.135841 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.135863 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.135876 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.233728 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:17 crc kubenswrapper[4788]: E1010 14:46:17.233854 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.238071 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.238106 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.238116 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.238129 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.238164 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.340851 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.340891 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.340900 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.340918 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.340928 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.443482 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.443524 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.443536 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.443552 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.443563 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.546064 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.546102 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.546110 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.546128 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.546154 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.648541 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.648576 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.648584 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.648597 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.648607 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.751065 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.751131 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.751164 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.751184 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.751199 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.853744 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.853787 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.853798 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.853814 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.853825 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.955819 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.955860 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.955871 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.955885 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:17 crc kubenswrapper[4788]: I1010 14:46:17.955893 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:17Z","lastTransitionTime":"2025-10-10T14:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.058418 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.058455 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.058463 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.058478 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.058486 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.161686 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.161736 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.161749 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.161767 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.161780 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.233912 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:18 crc kubenswrapper[4788]: E1010 14:46:18.234016 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.234066 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.234209 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:18 crc kubenswrapper[4788]: E1010 14:46:18.234302 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:18 crc kubenswrapper[4788]: E1010 14:46:18.234337 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.248616 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.260307 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.263667 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.263701 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.263709 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.263727 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.263738 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.272780 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.285769 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"2025-10-10T14:45:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44\\\\n2025-10-10T14:45:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44 to /host/opt/cni/bin/\\\\n2025-10-10T14:45:20Z [verbose] multus-daemon started\\\\n2025-10-10T14:45:20Z [verbose] Readiness Indicator file check\\\\n2025-10-10T14:46:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:46:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.298720 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.312540 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.327718 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.340612 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.359790 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.367687 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.367733 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.367746 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.367766 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.367778 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.376469 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.391057 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.401675 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.415748 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.426661 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.441396 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.463105 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:12Z\\\",\\\"message\\\":\\\"nsole]} name:Service_openshift-console/console_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.194:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d7d7b270-1480-47f8-bdf9-690dbab310cb}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1010 14:46:12.051320 6784 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nI1010 14:46:12.051326 6784 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nI1010 14:46:12.051330 6784 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx in node crc\\\\nI1010 14:46:12.051334 6784 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx after 0 failed attempt(s)\\\\nI1010 14:46:12.051338 6784 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nF1010 14:46:12.051352 6784 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:46:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.470638 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.470777 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.470792 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.470809 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.471122 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.475120 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.488280 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:18Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.573589 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.573618 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.573626 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.573638 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.573647 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.676303 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.676366 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.676383 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.676405 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.676420 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.777953 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.778012 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.778023 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.778037 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.778048 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.880189 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.880230 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.880242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.880256 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.880265 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.982679 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.982707 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.982715 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.982730 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:18 crc kubenswrapper[4788]: I1010 14:46:18.982739 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:18Z","lastTransitionTime":"2025-10-10T14:46:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.085245 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.085629 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.085791 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.086007 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.086209 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.189113 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.189173 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.189185 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.189200 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.189210 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.233297 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:19 crc kubenswrapper[4788]: E1010 14:46:19.233473 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.291787 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.291859 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.291869 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.291884 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.291893 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.395047 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.395195 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.395217 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.395243 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.395259 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.499133 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.499294 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.499336 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.499377 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.499399 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.602823 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.602902 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.602920 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.602954 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.602976 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.706605 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.706651 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.706665 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.706684 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.706698 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.809318 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.809377 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.809394 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.809421 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.809444 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.911534 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.911584 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.911595 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.911609 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:19 crc kubenswrapper[4788]: I1010 14:46:19.911618 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:19Z","lastTransitionTime":"2025-10-10T14:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.019100 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.019161 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.019172 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.019192 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.019202 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.121438 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.121507 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.121526 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.121556 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.121574 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.224323 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.224386 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.224401 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.224423 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.224438 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.233916 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.233921 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.234019 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:20 crc kubenswrapper[4788]: E1010 14:46:20.234243 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:20 crc kubenswrapper[4788]: E1010 14:46:20.234578 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:20 crc kubenswrapper[4788]: E1010 14:46:20.234841 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.326732 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.326762 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.326770 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.326783 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.326792 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.429809 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.429864 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.429879 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.429895 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.429904 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.533097 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.533150 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.533160 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.533174 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.533183 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.636605 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.637228 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.637469 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.637696 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.637878 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.741196 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.741278 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.741295 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.741768 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.741794 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.844465 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.844537 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.844557 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.844592 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.844613 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.948122 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.948235 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.948253 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.948281 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:20 crc kubenswrapper[4788]: I1010 14:46:20.948300 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:20Z","lastTransitionTime":"2025-10-10T14:46:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.050926 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.050986 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.050997 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.051012 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.051024 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.153633 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.153745 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.153770 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.153800 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.153821 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.233378 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.233510 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.256236 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.256272 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.256285 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.256299 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.256311 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.360024 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.360065 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.360076 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.360095 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.360107 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.463202 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.463251 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.463260 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.463277 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.463290 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.566048 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.566093 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.566113 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.566292 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.566321 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.669350 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.669438 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.669465 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.669497 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.669519 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.772497 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.772548 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.772559 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.772575 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.772590 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.875672 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.875771 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.875792 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.875828 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.875852 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.980177 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.980600 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.980618 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.980642 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.980658 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:21Z","lastTransitionTime":"2025-10-10T14:46:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.988975 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.989247 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.989273 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.98924245 +0000 UTC m=+148.438958048 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.989391 4788 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.989615 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.989587008 +0000 UTC m=+148.439302586 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.989832 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.989954 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.990061 4788 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.990210 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.990200274 +0000 UTC m=+148.439915812 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.989511 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:21 crc kubenswrapper[4788]: I1010 14:46:21.990415 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.990547 4788 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:46:21 crc kubenswrapper[4788]: E1010 14:46:21.990635 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.990613143 +0000 UTC m=+148.440328741 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.083712 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.083776 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.083802 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.083824 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.083840 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.091218 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:22 crc kubenswrapper[4788]: E1010 14:46:22.091561 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 10 14:46:22 crc kubenswrapper[4788]: E1010 14:46:22.091678 4788 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 10 14:46:22 crc kubenswrapper[4788]: E1010 14:46:22.091779 4788 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:46:22 crc kubenswrapper[4788]: E1010 14:46:22.091913 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.091896339 +0000 UTC m=+148.541611897 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.186937 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.187375 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.187536 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.187689 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.187823 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.233345 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.233469 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.233469 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:22 crc kubenswrapper[4788]: E1010 14:46:22.233846 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:22 crc kubenswrapper[4788]: E1010 14:46:22.234553 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:22 crc kubenswrapper[4788]: E1010 14:46:22.234780 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.290988 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.291053 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.291067 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.291087 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.291100 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.395668 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.396031 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.396185 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.396345 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.396503 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.499647 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.499694 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.499703 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.499720 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.499729 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.604215 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.604296 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.604324 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.604354 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.604374 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.707787 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.707837 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.707850 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.707871 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.707884 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.810318 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.810362 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.810371 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.810391 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.810402 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.916783 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.916822 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.916830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.916845 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:22 crc kubenswrapper[4788]: I1010 14:46:22.916854 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:22Z","lastTransitionTime":"2025-10-10T14:46:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.019752 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.019798 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.019808 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.019823 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.019832 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.122782 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.122874 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.122891 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.122909 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.122921 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.226035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.226095 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.226113 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.226136 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.226183 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.233731 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:23 crc kubenswrapper[4788]: E1010 14:46:23.233895 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.329440 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.329555 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.329574 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.329597 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.329615 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.432919 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.432990 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.433011 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.433041 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.433062 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.536278 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.536360 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.536384 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.536414 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.536437 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.640350 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.640435 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.640460 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.640493 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.640517 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.745477 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.745595 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.745614 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.745639 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.745655 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.848885 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.848945 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.848969 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.848998 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.849022 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.952717 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.952789 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.952814 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.952848 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:23 crc kubenswrapper[4788]: I1010 14:46:23.952879 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:23Z","lastTransitionTime":"2025-10-10T14:46:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.056407 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.056616 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.056652 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.056735 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.056802 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.160401 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.160467 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.160486 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.160512 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.160529 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.232869 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:24 crc kubenswrapper[4788]: E1010 14:46:24.233044 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.232888 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.233244 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:24 crc kubenswrapper[4788]: E1010 14:46:24.233261 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:24 crc kubenswrapper[4788]: E1010 14:46:24.233457 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.264057 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.264434 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.264558 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.264677 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.264797 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.368661 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.369028 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.369119 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.369260 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.369343 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.472460 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.472520 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.472537 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.472564 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.472582 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.575207 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.575673 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.575818 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.575941 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.576068 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.679342 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.679422 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.679439 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.679458 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.679509 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.782100 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.782147 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.782156 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.782170 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.782180 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.887204 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.887256 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.887271 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.887291 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.887310 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.989251 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.989556 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.989626 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.989725 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:24 crc kubenswrapper[4788]: I1010 14:46:24.989795 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:24Z","lastTransitionTime":"2025-10-10T14:46:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.091937 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.091975 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.091986 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.092002 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.092013 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.194925 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.194992 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.195009 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.195035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.195054 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.233455 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:25 crc kubenswrapper[4788]: E1010 14:46:25.233628 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.297382 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.297452 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.297471 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.297496 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.297513 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.400126 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.400212 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.400228 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.400247 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.400263 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.502918 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.502964 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.502975 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.502994 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.503006 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.605290 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.605324 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.605334 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.605349 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.605360 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.707813 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.707865 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.707883 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.707907 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.707926 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.810239 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.810303 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.810318 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.810335 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.810347 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.913297 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.913360 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.913380 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.913408 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:25 crc kubenswrapper[4788]: I1010 14:46:25.913431 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:25Z","lastTransitionTime":"2025-10-10T14:46:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.015929 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.015978 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.015996 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.016018 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.016034 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.119515 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.119559 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.119569 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.119587 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.119596 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.125839 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.125921 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.125948 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.125978 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.126011 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.149332 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.153302 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.153369 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.153388 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.153418 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.153441 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.172622 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.177132 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.177253 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.177317 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.177385 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.177448 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.190111 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.195614 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.195666 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.195682 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.195704 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.195718 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.209199 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.213009 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.213125 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.213202 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.213292 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.213363 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.225363 4788 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f6a8a518-4f69-41df-a3e7-d3eaeece5119\\\",\\\"systemUUID\\\":\\\"e666a902-fb0b-4061-82b8-ccc08d87745f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:26Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.225513 4788 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.227079 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.227121 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.227161 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.227187 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.227234 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.233597 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.233660 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.233734 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.233804 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.233752 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:26 crc kubenswrapper[4788]: E1010 14:46:26.234013 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.329670 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.329720 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.329733 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.329750 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.329763 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.431839 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.431874 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.431884 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.431899 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.431910 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.535806 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.535856 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.535870 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.535886 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.535895 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.638910 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.638953 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.638966 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.638985 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.639000 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.741944 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.741990 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.742003 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.742021 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.742033 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.845034 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.845093 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.845111 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.845134 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.845218 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.951106 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.951193 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.951215 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.951244 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:26 crc kubenswrapper[4788]: I1010 14:46:26.951267 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:26Z","lastTransitionTime":"2025-10-10T14:46:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.054484 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.054808 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.054927 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.055064 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.055232 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.158589 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.158639 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.158654 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.158677 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.158690 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.233423 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:27 crc kubenswrapper[4788]: E1010 14:46:27.233624 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.234442 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:46:27 crc kubenswrapper[4788]: E1010 14:46:27.234631 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.260564 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.260615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.260630 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.260650 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.260664 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.363886 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.363952 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.363967 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.363987 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.363999 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.466226 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.466281 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.466292 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.466311 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.466328 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.569287 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.569623 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.569755 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.569853 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.569953 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.672618 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.673183 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.673564 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.673781 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.673954 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.776800 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.776864 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.776882 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.776910 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.776930 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.880464 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.880521 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.880538 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.880565 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.880582 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.984737 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.984807 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.984830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.984860 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:27 crc kubenswrapper[4788]: I1010 14:46:27.984882 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:27Z","lastTransitionTime":"2025-10-10T14:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.088498 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.088544 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.088556 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.088572 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.088584 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.191294 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.191417 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.191440 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.191469 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.191486 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.233628 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:28 crc kubenswrapper[4788]: E1010 14:46:28.233738 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.233913 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:28 crc kubenswrapper[4788]: E1010 14:46:28.234005 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.234183 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:28 crc kubenswrapper[4788]: E1010 14:46:28.234403 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.256869 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17c0189a-e025-412b-a0b1-c649daddc41b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://beb436294e69cb527afde418b8a4f0c92e8c35ec70a14fa4a9a740f1d90708dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80665af1ce3d25a005615be05097be86fdc5a1d8da91313b34bfb3f45a146654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e8aeb9786e5e176f42d3674306ff452b1a5ee4e7e89d01dca87d6dfe26c8974\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://720d58d28044f9663a4ede7444aaddadd366aae562be7478a63a7472e97f370c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de769c98cc41e4aa62720fcc2827ca1e13489d11d7aab3440a4d23d19b4b859\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ec4c442c8a66dc3a5b057758655874ef6880af70f2e6057c6d6f2e380aaf27d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e604e789ff7fff1aaa1dd7490f550ab9932ebc03ec94358e61df1b478a14a6e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4daad7d1f60d64adfc7d03216449383e6aefd6c7afed9a9e7769f355103b84e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.278375 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ff8f78b-529a-41ff-949f-4ea49461f9f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19747d7b18c954a3ad2df84597820f27d6047f791c08a6c55b09c534dd8a9082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d44ec431812495ed5dbbd9b490d27d62c43958f71c39d7789bbe54559938c8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6e1efa6731adbebdb1e98054242f2b4802d7a55cec8831da9627ecf9e0cdf03\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c00bfdd43c897c965ed9700c5a8277cce77f7e971201234a227fa3866bb0dd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.289837 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2pg5g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f399508-57f4-4ee0-9aa1-dc6cf9b43d2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ecb21c9c62ad32ba4daa84f9c17debd1a73c2ace1c82497ae039c2f64f5a135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-klhwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:22Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2pg5g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.294017 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.294101 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.294118 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.294151 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.294165 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.304077 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50ef40141a518ff51b9057e7816ae956c09cbec3010f7910b58c6ab68fcc94b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.322226 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-szmv7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c8cb468-2b03-45c9-93e9-abbcc4c075a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91ed10710f17266cf7bcccb27864c60509ab7eeb290c777a0b9e0dd519ac7cd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ce1f10d3dc3f34573cf39040702a53c6b789cf5abfa7e7291dc572bfb9c2182\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6570cde239ad7ae5d877f0d8a1288a52505a90c408e1bce420042d4ff3051a4d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72134134c25020961289a4516ddc9d7cbbe1935e56bdfe00875b86172f4b85c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://171168eb1a9337ccd7044877656d399703b0300e69d82a6978c3b7cf28a165cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ca4a2f9664f2d23d007d4f2e5b50b625951a8b894709689b657631de351e361\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65e8d90f9b0dfaa840c97ba1ae47445ebad9d6d12a785da3ac7dcd514da787e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fnvfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-szmv7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.342763 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2419d2de-214a-4a13-b941-2acd571f0360\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac23ee2bfe7ef168ce9e10edf77b0ed805cc1ca446e311457ab5f371cc90c91b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rh4sp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dxfdf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.356508 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fff2dd72-c73c-455d-baa0-5f71322f1c0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://789b4633eec5373afd3984ca6b8de520d22e62a7bdc0d5473d894fd4463b4281\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://761ca54206fe543b982f0dbc25a78a443118af827d7926d5bd7bada3a95b1efb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g58j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-448bx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.367272 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3d61e9aa-0688-436b-a8af-d6d8353c4350\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jxgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:31Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jqlwd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.387074 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4c30755-ef02-49eb-96d2-2214c14eaab9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7c2fda15e8daa9cd173edd7590a798463de0c8c39869c5282df997966b5ce4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d01250703faaf5027cd0bdbb80f67ef021689568532c2efc939cf982575a468\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c62cdd681385ff12c1b8b4717721eb9db79b8cbd4ecb691d937e6a0f8a50bf5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3851e51d7ffe7122c8d83fe083148e0bcaffc471442f3b9707189d678203f95\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://435cd265e6ba986e2c5e8cd7f2f69e092aeb0abc0bf63b8d11b1ec3f89224c63\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-10T14:45:17Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1010 14:45:11.797501 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1010 14:45:11.800158 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-500215970/tls.crt::/tmp/serving-cert-500215970/tls.key\\\\\\\"\\\\nI1010 14:45:17.334408 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1010 14:45:17.339491 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1010 14:45:17.339523 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1010 14:45:17.339548 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1010 14:45:17.339554 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1010 14:45:17.347918 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1010 14:45:17.347939 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1010 14:45:17.347985 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.347996 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1010 14:45:17.348002 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1010 14:45:17.348008 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1010 14:45:17.348012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1010 14:45:17.348016 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1010 14:45:17.350380 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e7d757fab17af1a465c2238edba620478d42c55589f37f63ef96689e8f4e497\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6a73c09fe4bffd1ba54fd59df1ffa1649aac53015be54d0c4fe81a114364ebc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.398009 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.398057 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.398067 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.398095 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.398107 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.403629 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.429731 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"baeca99f-afc8-4129-8110-51e79dd8b054\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:12Z\\\",\\\"message\\\":\\\"nsole]} name:Service_openshift-console/console_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.194:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d7d7b270-1480-47f8-bdf9-690dbab310cb}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1010 14:46:12.051320 6784 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nI1010 14:46:12.051326 6784 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nI1010 14:46:12.051330 6784 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx in node crc\\\\nI1010 14:46:12.051334 6784 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx after 0 failed attempt(s)\\\\nI1010 14:46:12.051338 6784 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx\\\\nF1010 14:46:12.051352 6784 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:46:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kckf4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cwgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.446430 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fdhpt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"027f048a-2237-4fa6-9bc5-4bd981c13b7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0f3b3ee17365d136fab36a9b03a3a0f91f9d696287a4f9ceb6538ba1bba3249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvwpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fdhpt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.460740 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7f2a5dc2-cd46-438f-80c4-e532bb6b0ae8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:44:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1f584a47db95b27d82113c8bc78e64b084cdd33752a18a8fcba4b99b808d09c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14dcfefada79be50bee27a1e33143137a3c16ad12ab291c2c58680453ba7ba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71c4d9410f5f9db7e809fddddf8140642731300812ce109650d79043d580a895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://125a887d67500a7b6cfe9d7eeb2108afa5d0caf72f41eae73e096b08f55c24e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-10T14:44:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-10T14:44:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:44:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.474470 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.495615 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09e2f443df7b17a509a9c87dd94003f5b7cbc3c7de92233728d1ec3eba87861a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.502748 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.503066 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.503304 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.503502 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.503690 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.509947 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48075c418a1f75b015fe67e3c98fa0a4456ed9bc47151cca05a40944d473f86f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34f6f39fec69b5375a99bf607312723737be832b45287ab82b75116ead34f61e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.525550 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rzgfk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39cfa1ec-4912-4550-9ead-7f6113db2221\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-10T14:46:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-10T14:46:05Z\\\",\\\"message\\\":\\\"2025-10-10T14:45:20+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44\\\\n2025-10-10T14:45:20+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9b3861dd-885e-4cf0-a383-fef0fff08a44 to /host/opt/cni/bin/\\\\n2025-10-10T14:45:20Z [verbose] multus-daemon started\\\\n2025-10-10T14:45:20Z [verbose] Readiness Indicator file check\\\\n2025-10-10T14:46:05Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-10T14:45:18Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-10T14:46:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fvlx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-10T14:45:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rzgfk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.538483 4788 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-10T14:45:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-10T14:46:28Z is after 2025-08-24T17:21:41Z" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.607951 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.608015 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.608035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.608063 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.608085 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.710841 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.710879 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.710888 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.710903 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.710913 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.812714 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.812755 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.812764 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.812777 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.812786 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.914836 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.914909 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.914930 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.914959 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:28 crc kubenswrapper[4788]: I1010 14:46:28.914983 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:28Z","lastTransitionTime":"2025-10-10T14:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.017335 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.017378 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.017387 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.017402 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.017411 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.120092 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.120149 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.120159 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.120176 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.120184 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.222907 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.222974 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.222987 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.223007 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.223022 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.233245 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:29 crc kubenswrapper[4788]: E1010 14:46:29.233405 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.324995 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.325035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.325044 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.325062 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.325073 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.428404 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.428485 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.428510 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.428546 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.428571 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.531585 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.531678 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.531697 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.531722 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.531739 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.634970 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.635022 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.635038 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.635062 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.635079 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.737259 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.737302 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.737311 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.737328 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.737337 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.839676 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.839722 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.839732 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.839749 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.839758 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.942723 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.942768 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.942781 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.942799 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:29 crc kubenswrapper[4788]: I1010 14:46:29.942811 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:29Z","lastTransitionTime":"2025-10-10T14:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.045359 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.045430 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.045441 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.045472 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.045483 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.147615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.147709 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.147745 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.147776 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.147801 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.233700 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:30 crc kubenswrapper[4788]: E1010 14:46:30.233835 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.233724 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:30 crc kubenswrapper[4788]: E1010 14:46:30.234049 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.234360 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:30 crc kubenswrapper[4788]: E1010 14:46:30.234554 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.250183 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.250497 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.250575 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.250647 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.250719 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.353978 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.354062 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.354078 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.354096 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.354110 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.456617 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.456668 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.456683 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.456705 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.456720 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.562123 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.562175 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.562184 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.562201 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.562211 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.665073 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.665124 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.665132 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.665165 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.665175 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.767839 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.767885 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.767895 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.767910 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.767921 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.870795 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.870846 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.870857 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.870872 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.871228 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.973710 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.973738 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.973745 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.973757 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:30 crc kubenswrapper[4788]: I1010 14:46:30.973765 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:30Z","lastTransitionTime":"2025-10-10T14:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.075997 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.076043 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.076060 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.076076 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.076085 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.179658 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.179730 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.179749 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.179777 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.179797 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.233592 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:31 crc kubenswrapper[4788]: E1010 14:46:31.233866 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.247268 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.283105 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.283211 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.283242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.283274 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.283295 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.386464 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.386561 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.386583 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.386617 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.386640 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.490449 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.490516 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.490535 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.490566 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.490585 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.592704 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.592745 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.592757 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.592774 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.592785 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.696074 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.696197 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.696225 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.696259 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.696283 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.798837 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.798908 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.798930 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.798957 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.798981 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.902002 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.902060 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.902081 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.902111 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:31 crc kubenswrapper[4788]: I1010 14:46:31.902134 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:31Z","lastTransitionTime":"2025-10-10T14:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.004287 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.004333 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.004348 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.004372 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.004386 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.107543 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.107600 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.107615 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.107641 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.107657 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.210357 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.210393 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.210402 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.210414 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.210423 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.233126 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.233210 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.233130 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:32 crc kubenswrapper[4788]: E1010 14:46:32.233306 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:32 crc kubenswrapper[4788]: E1010 14:46:32.233464 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:32 crc kubenswrapper[4788]: E1010 14:46:32.233586 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.312245 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.312291 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.312303 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.312320 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.312333 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.418826 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.418882 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.418902 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.418921 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.418938 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.521830 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.521869 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.521879 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.521891 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.521900 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.624401 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.624783 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.624862 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.624940 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.625030 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.727910 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.727967 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.727977 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.727995 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.728005 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.830691 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.830749 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.830760 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.830774 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.830799 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.932979 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.933008 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.933038 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.933054 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:32 crc kubenswrapper[4788]: I1010 14:46:32.933064 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:32Z","lastTransitionTime":"2025-10-10T14:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.035208 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.035246 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.035256 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.035271 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.035282 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.137576 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.137659 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.137674 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.137693 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.137706 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.233576 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:33 crc kubenswrapper[4788]: E1010 14:46:33.233695 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.239659 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.239691 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.239700 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.239714 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.239724 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.341470 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.341764 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.341843 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.341940 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.342024 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.444918 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.445596 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.445608 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.445621 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.445629 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.547438 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.547478 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.547488 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.547504 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.547515 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.649653 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.649693 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.649703 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.649717 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.649726 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.754295 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.754394 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.754422 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.754449 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.754470 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.856944 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.856992 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.857010 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.857034 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.857049 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.959035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.959078 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.959087 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.959101 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:33 crc kubenswrapper[4788]: I1010 14:46:33.959110 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:33Z","lastTransitionTime":"2025-10-10T14:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.061596 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.061629 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.061640 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.061653 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.061662 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.164425 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.164461 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.164470 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.164486 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.164496 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.233332 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:34 crc kubenswrapper[4788]: E1010 14:46:34.233472 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.233557 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.233336 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:34 crc kubenswrapper[4788]: E1010 14:46:34.233662 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:34 crc kubenswrapper[4788]: E1010 14:46:34.233809 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.267304 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.267362 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.267369 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.267384 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.267394 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.369360 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.369417 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.369431 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.369449 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.369461 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.471165 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.471208 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.471218 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.471234 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.471246 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.573098 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.573158 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.573170 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.573188 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.573199 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.676002 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.676041 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.676061 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.676076 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.676086 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.777991 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.778027 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.778035 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.778048 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.778059 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.880313 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.880359 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.880371 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.880387 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.880398 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.982214 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.982254 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.982262 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.982278 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:34 crc kubenswrapper[4788]: I1010 14:46:34.982286 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:34Z","lastTransitionTime":"2025-10-10T14:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.084716 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.084995 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.085131 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.085255 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.085354 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.187804 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.188178 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.188305 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.188466 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.188567 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.233426 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:35 crc kubenswrapper[4788]: E1010 14:46:35.233556 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.291559 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.291594 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.291605 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.291621 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.291630 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.393971 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.394018 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.394026 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.394041 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.394050 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.496097 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.496161 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.496171 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.496185 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.496195 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.598301 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.598354 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.598368 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.598386 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.598400 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.700988 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.701030 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.701041 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.701055 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.701065 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.803510 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.803558 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.803569 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.803586 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.803597 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.832713 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:35 crc kubenswrapper[4788]: E1010 14:46:35.832824 4788 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:46:35 crc kubenswrapper[4788]: E1010 14:46:35.832873 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs podName:3d61e9aa-0688-436b-a8af-d6d8353c4350 nodeName:}" failed. No retries permitted until 2025-10-10 14:47:39.832858741 +0000 UTC m=+162.282574289 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs") pod "network-metrics-daemon-jqlwd" (UID: "3d61e9aa-0688-436b-a8af-d6d8353c4350") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.906116 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.906204 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.906220 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.906242 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:35 crc kubenswrapper[4788]: I1010 14:46:35.906259 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:35Z","lastTransitionTime":"2025-10-10T14:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.008804 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.008835 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.008844 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.008856 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.008864 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:36Z","lastTransitionTime":"2025-10-10T14:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.111686 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.111738 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.111748 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.111766 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.112115 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:36Z","lastTransitionTime":"2025-10-10T14:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.214596 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.214634 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.214643 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.214659 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.214669 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:36Z","lastTransitionTime":"2025-10-10T14:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.233312 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.233349 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.233798 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:36 crc kubenswrapper[4788]: E1010 14:46:36.233891 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:36 crc kubenswrapper[4788]: E1010 14:46:36.234200 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:36 crc kubenswrapper[4788]: E1010 14:46:36.234276 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.316604 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.316644 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.316653 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.316668 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.316677 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:36Z","lastTransitionTime":"2025-10-10T14:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.418768 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.418809 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.418820 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.418842 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.418854 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:36Z","lastTransitionTime":"2025-10-10T14:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.521511 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.521561 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.521570 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.521587 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.521597 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:36Z","lastTransitionTime":"2025-10-10T14:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.552522 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.552572 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.552585 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.552607 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.552628 4788 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-10T14:46:36Z","lastTransitionTime":"2025-10-10T14:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.608708 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr"] Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.609069 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.610815 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.610876 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.611794 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.613795 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.637950 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b7d30350-4c2b-459e-9b2e-fc72b38dc505-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.638364 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b7d30350-4c2b-459e-9b2e-fc72b38dc505-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.638406 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7d30350-4c2b-459e-9b2e-fc72b38dc505-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.638447 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7d30350-4c2b-459e-9b2e-fc72b38dc505-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.638468 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7d30350-4c2b-459e-9b2e-fc72b38dc505-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.651290 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.651269215 podStartE2EDuration="48.651269215s" podCreationTimestamp="2025-10-10 14:45:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.651175013 +0000 UTC m=+99.100890571" watchObservedRunningTime="2025-10-10 14:46:36.651269215 +0000 UTC m=+99.100984763" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.735707 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-rzgfk" podStartSLOduration=79.73568617 podStartE2EDuration="1m19.73568617s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.709927893 +0000 UTC m=+99.159643441" watchObservedRunningTime="2025-10-10 14:46:36.73568617 +0000 UTC m=+99.185401718" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.739262 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b7d30350-4c2b-459e-9b2e-fc72b38dc505-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.739304 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b7d30350-4c2b-459e-9b2e-fc72b38dc505-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.739333 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7d30350-4c2b-459e-9b2e-fc72b38dc505-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.739367 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7d30350-4c2b-459e-9b2e-fc72b38dc505-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.739387 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7d30350-4c2b-459e-9b2e-fc72b38dc505-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.739480 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b7d30350-4c2b-459e-9b2e-fc72b38dc505-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.739554 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b7d30350-4c2b-459e-9b2e-fc72b38dc505-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.741065 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b7d30350-4c2b-459e-9b2e-fc72b38dc505-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.744929 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7d30350-4c2b-459e-9b2e-fc72b38dc505-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.752907 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=74.752888798 podStartE2EDuration="1m14.752888798s" podCreationTimestamp="2025-10-10 14:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.751651078 +0000 UTC m=+99.201366626" watchObservedRunningTime="2025-10-10 14:46:36.752888798 +0000 UTC m=+99.202604346" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.753840 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=76.753832111 podStartE2EDuration="1m16.753832111s" podCreationTimestamp="2025-10-10 14:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.736659994 +0000 UTC m=+99.186375562" watchObservedRunningTime="2025-10-10 14:46:36.753832111 +0000 UTC m=+99.203547659" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.760184 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b7d30350-4c2b-459e-9b2e-fc72b38dc505-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lv5tr\" (UID: \"b7d30350-4c2b-459e-9b2e-fc72b38dc505\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.761508 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2pg5g" podStartSLOduration=78.761493287 podStartE2EDuration="1m18.761493287s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.760537045 +0000 UTC m=+99.210252593" watchObservedRunningTime="2025-10-10 14:46:36.761493287 +0000 UTC m=+99.211208835" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.793016 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=5.792997535 podStartE2EDuration="5.792997535s" podCreationTimestamp="2025-10-10 14:46:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.780646774 +0000 UTC m=+99.230362332" watchObservedRunningTime="2025-10-10 14:46:36.792997535 +0000 UTC m=+99.242713083" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.817683 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-szmv7" podStartSLOduration=79.817655546 podStartE2EDuration="1m19.817655546s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.807395465 +0000 UTC m=+99.257111013" watchObservedRunningTime="2025-10-10 14:46:36.817655546 +0000 UTC m=+99.267371104" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.831865 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podStartSLOduration=79.831848371 podStartE2EDuration="1m19.831848371s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.818857764 +0000 UTC m=+99.268573322" watchObservedRunningTime="2025-10-10 14:46:36.831848371 +0000 UTC m=+99.281563919" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.832024 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-448bx" podStartSLOduration=78.832021385 podStartE2EDuration="1m18.832021385s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.831966364 +0000 UTC m=+99.281681912" watchObservedRunningTime="2025-10-10 14:46:36.832021385 +0000 UTC m=+99.281736933" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.858265 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.858248364 podStartE2EDuration="1m18.858248364s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.858070079 +0000 UTC m=+99.307785627" watchObservedRunningTime="2025-10-10 14:46:36.858248364 +0000 UTC m=+99.307963912" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.928980 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" Oct 10 14:46:36 crc kubenswrapper[4788]: I1010 14:46:36.933126 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-fdhpt" podStartSLOduration=79.933114336 podStartE2EDuration="1m19.933114336s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:36.93244408 +0000 UTC m=+99.382159628" watchObservedRunningTime="2025-10-10 14:46:36.933114336 +0000 UTC m=+99.382829884" Oct 10 14:46:37 crc kubenswrapper[4788]: I1010 14:46:37.232804 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:37 crc kubenswrapper[4788]: E1010 14:46:37.232984 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:37 crc kubenswrapper[4788]: I1010 14:46:37.831366 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" event={"ID":"b7d30350-4c2b-459e-9b2e-fc72b38dc505","Type":"ContainerStarted","Data":"f42bcd5070a2c0cd5aeffe77da6f4816da1a3d9ae0b662f11ca6e236e2b73068"} Oct 10 14:46:37 crc kubenswrapper[4788]: I1010 14:46:37.831417 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" event={"ID":"b7d30350-4c2b-459e-9b2e-fc72b38dc505","Type":"ContainerStarted","Data":"cfda8d7ebd6d167e143a917c33dda0eabad1b402c6fbbae9e6083a5f25bbb677"} Oct 10 14:46:37 crc kubenswrapper[4788]: I1010 14:46:37.847724 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lv5tr" podStartSLOduration=80.84769445 podStartE2EDuration="1m20.84769445s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:37.847066815 +0000 UTC m=+100.296782373" watchObservedRunningTime="2025-10-10 14:46:37.84769445 +0000 UTC m=+100.297410008" Oct 10 14:46:38 crc kubenswrapper[4788]: I1010 14:46:38.233401 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:38 crc kubenswrapper[4788]: I1010 14:46:38.233568 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:38 crc kubenswrapper[4788]: E1010 14:46:38.234576 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:38 crc kubenswrapper[4788]: I1010 14:46:38.234605 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:38 crc kubenswrapper[4788]: E1010 14:46:38.234812 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:38 crc kubenswrapper[4788]: E1010 14:46:38.234903 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:39 crc kubenswrapper[4788]: I1010 14:46:39.233581 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:39 crc kubenswrapper[4788]: E1010 14:46:39.233835 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:40 crc kubenswrapper[4788]: I1010 14:46:40.235085 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:46:40 crc kubenswrapper[4788]: I1010 14:46:40.235354 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:40 crc kubenswrapper[4788]: I1010 14:46:40.235384 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:40 crc kubenswrapper[4788]: I1010 14:46:40.235430 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:40 crc kubenswrapper[4788]: E1010 14:46:40.235489 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:40 crc kubenswrapper[4788]: E1010 14:46:40.235631 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:40 crc kubenswrapper[4788]: E1010 14:46:40.235722 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cwgns_openshift-ovn-kubernetes(baeca99f-afc8-4129-8110-51e79dd8b054)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" Oct 10 14:46:40 crc kubenswrapper[4788]: E1010 14:46:40.235797 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:41 crc kubenswrapper[4788]: I1010 14:46:41.233715 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:41 crc kubenswrapper[4788]: E1010 14:46:41.233887 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:42 crc kubenswrapper[4788]: I1010 14:46:42.233341 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:42 crc kubenswrapper[4788]: E1010 14:46:42.233464 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:42 crc kubenswrapper[4788]: I1010 14:46:42.233538 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:42 crc kubenswrapper[4788]: I1010 14:46:42.233603 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:42 crc kubenswrapper[4788]: E1010 14:46:42.233715 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:42 crc kubenswrapper[4788]: E1010 14:46:42.233865 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:43 crc kubenswrapper[4788]: I1010 14:46:43.232977 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:43 crc kubenswrapper[4788]: E1010 14:46:43.233410 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:44 crc kubenswrapper[4788]: I1010 14:46:44.233623 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:44 crc kubenswrapper[4788]: E1010 14:46:44.233758 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:44 crc kubenswrapper[4788]: I1010 14:46:44.233828 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:44 crc kubenswrapper[4788]: I1010 14:46:44.233944 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:44 crc kubenswrapper[4788]: E1010 14:46:44.233980 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:44 crc kubenswrapper[4788]: E1010 14:46:44.234003 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:45 crc kubenswrapper[4788]: I1010 14:46:45.232768 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:45 crc kubenswrapper[4788]: E1010 14:46:45.232901 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:46 crc kubenswrapper[4788]: I1010 14:46:46.233814 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:46 crc kubenswrapper[4788]: I1010 14:46:46.233807 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:46 crc kubenswrapper[4788]: E1010 14:46:46.234061 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:46 crc kubenswrapper[4788]: E1010 14:46:46.234235 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:46 crc kubenswrapper[4788]: I1010 14:46:46.235505 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:46 crc kubenswrapper[4788]: E1010 14:46:46.235898 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:47 crc kubenswrapper[4788]: I1010 14:46:47.233587 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:47 crc kubenswrapper[4788]: E1010 14:46:47.234043 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:48 crc kubenswrapper[4788]: I1010 14:46:48.233116 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:48 crc kubenswrapper[4788]: I1010 14:46:48.233195 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:48 crc kubenswrapper[4788]: I1010 14:46:48.233133 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:48 crc kubenswrapper[4788]: E1010 14:46:48.235127 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:48 crc kubenswrapper[4788]: E1010 14:46:48.235258 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:48 crc kubenswrapper[4788]: E1010 14:46:48.235407 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:49 crc kubenswrapper[4788]: I1010 14:46:49.233118 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:49 crc kubenswrapper[4788]: E1010 14:46:49.233340 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:50 crc kubenswrapper[4788]: I1010 14:46:50.233409 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:50 crc kubenswrapper[4788]: I1010 14:46:50.233457 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:50 crc kubenswrapper[4788]: I1010 14:46:50.233483 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:50 crc kubenswrapper[4788]: E1010 14:46:50.233580 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:50 crc kubenswrapper[4788]: E1010 14:46:50.233638 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:50 crc kubenswrapper[4788]: E1010 14:46:50.233729 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:51 crc kubenswrapper[4788]: I1010 14:46:51.233072 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:51 crc kubenswrapper[4788]: E1010 14:46:51.233473 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.233192 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.233222 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.233296 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:52 crc kubenswrapper[4788]: E1010 14:46:52.233361 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:52 crc kubenswrapper[4788]: E1010 14:46:52.233459 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:52 crc kubenswrapper[4788]: E1010 14:46:52.233808 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.234034 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.884350 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/3.log" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.887123 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerStarted","Data":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.887604 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.889344 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/1.log" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.889880 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/0.log" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.889935 4788 generic.go:334] "Generic (PLEG): container finished" podID="39cfa1ec-4912-4550-9ead-7f6113db2221" containerID="a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be" exitCode=1 Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.889965 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rzgfk" event={"ID":"39cfa1ec-4912-4550-9ead-7f6113db2221","Type":"ContainerDied","Data":"a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be"} Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.889999 4788 scope.go:117] "RemoveContainer" containerID="7f4b7a2dc0a7d9ec4bba8a4f44745e76c214ac97b98108fff5126afe8282c0d5" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.890366 4788 scope.go:117] "RemoveContainer" containerID="a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be" Oct 10 14:46:52 crc kubenswrapper[4788]: E1010 14:46:52.890508 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-rzgfk_openshift-multus(39cfa1ec-4912-4550-9ead-7f6113db2221)\"" pod="openshift-multus/multus-rzgfk" podUID="39cfa1ec-4912-4550-9ead-7f6113db2221" Oct 10 14:46:52 crc kubenswrapper[4788]: I1010 14:46:52.914508 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podStartSLOduration=95.914488748 podStartE2EDuration="1m35.914488748s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:46:52.913317699 +0000 UTC m=+115.363033267" watchObservedRunningTime="2025-10-10 14:46:52.914488748 +0000 UTC m=+115.364204296" Oct 10 14:46:53 crc kubenswrapper[4788]: I1010 14:46:53.037438 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jqlwd"] Oct 10 14:46:53 crc kubenswrapper[4788]: I1010 14:46:53.037557 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:53 crc kubenswrapper[4788]: E1010 14:46:53.037685 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:53 crc kubenswrapper[4788]: I1010 14:46:53.894585 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/1.log" Oct 10 14:46:54 crc kubenswrapper[4788]: I1010 14:46:54.232814 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:54 crc kubenswrapper[4788]: E1010 14:46:54.233458 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:54 crc kubenswrapper[4788]: I1010 14:46:54.233642 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:54 crc kubenswrapper[4788]: I1010 14:46:54.233725 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:54 crc kubenswrapper[4788]: I1010 14:46:54.233768 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:54 crc kubenswrapper[4788]: E1010 14:46:54.233936 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:54 crc kubenswrapper[4788]: E1010 14:46:54.234082 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:54 crc kubenswrapper[4788]: E1010 14:46:54.234272 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:56 crc kubenswrapper[4788]: I1010 14:46:56.233670 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:56 crc kubenswrapper[4788]: I1010 14:46:56.233753 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:56 crc kubenswrapper[4788]: E1010 14:46:56.233798 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:56 crc kubenswrapper[4788]: I1010 14:46:56.233798 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:56 crc kubenswrapper[4788]: I1010 14:46:56.233755 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:56 crc kubenswrapper[4788]: E1010 14:46:56.233865 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:56 crc kubenswrapper[4788]: E1010 14:46:56.233913 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:56 crc kubenswrapper[4788]: E1010 14:46:56.234070 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:58 crc kubenswrapper[4788]: E1010 14:46:58.179561 4788 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 10 14:46:58 crc kubenswrapper[4788]: I1010 14:46:58.233119 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:46:58 crc kubenswrapper[4788]: I1010 14:46:58.233169 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:46:58 crc kubenswrapper[4788]: I1010 14:46:58.233266 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:46:58 crc kubenswrapper[4788]: I1010 14:46:58.233283 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:46:58 crc kubenswrapper[4788]: E1010 14:46:58.234124 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:46:58 crc kubenswrapper[4788]: E1010 14:46:58.234311 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:46:58 crc kubenswrapper[4788]: E1010 14:46:58.234462 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:46:58 crc kubenswrapper[4788]: E1010 14:46:58.234576 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:46:58 crc kubenswrapper[4788]: E1010 14:46:58.338800 4788 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 10 14:47:00 crc kubenswrapper[4788]: I1010 14:47:00.233824 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:00 crc kubenswrapper[4788]: I1010 14:47:00.233858 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:00 crc kubenswrapper[4788]: I1010 14:47:00.234031 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:00 crc kubenswrapper[4788]: I1010 14:47:00.234006 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:00 crc kubenswrapper[4788]: E1010 14:47:00.234110 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:47:00 crc kubenswrapper[4788]: E1010 14:47:00.234213 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:47:00 crc kubenswrapper[4788]: E1010 14:47:00.234341 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:47:00 crc kubenswrapper[4788]: E1010 14:47:00.234595 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:47:02 crc kubenswrapper[4788]: I1010 14:47:02.233678 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:02 crc kubenswrapper[4788]: I1010 14:47:02.233733 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:02 crc kubenswrapper[4788]: I1010 14:47:02.233793 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:02 crc kubenswrapper[4788]: E1010 14:47:02.233881 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:47:02 crc kubenswrapper[4788]: I1010 14:47:02.234123 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:02 crc kubenswrapper[4788]: E1010 14:47:02.234234 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:47:02 crc kubenswrapper[4788]: E1010 14:47:02.234413 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:47:02 crc kubenswrapper[4788]: E1010 14:47:02.234733 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:47:03 crc kubenswrapper[4788]: E1010 14:47:03.340718 4788 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 10 14:47:04 crc kubenswrapper[4788]: I1010 14:47:04.232948 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:04 crc kubenswrapper[4788]: I1010 14:47:04.233009 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:04 crc kubenswrapper[4788]: I1010 14:47:04.233102 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:04 crc kubenswrapper[4788]: I1010 14:47:04.233396 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:04 crc kubenswrapper[4788]: E1010 14:47:04.233627 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:47:04 crc kubenswrapper[4788]: E1010 14:47:04.233737 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:47:04 crc kubenswrapper[4788]: E1010 14:47:04.233860 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:47:04 crc kubenswrapper[4788]: E1010 14:47:04.233993 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:47:06 crc kubenswrapper[4788]: I1010 14:47:06.233306 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:06 crc kubenswrapper[4788]: I1010 14:47:06.233389 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:06 crc kubenswrapper[4788]: I1010 14:47:06.233402 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:06 crc kubenswrapper[4788]: E1010 14:47:06.233474 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:47:06 crc kubenswrapper[4788]: E1010 14:47:06.233576 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:47:06 crc kubenswrapper[4788]: I1010 14:47:06.233688 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:06 crc kubenswrapper[4788]: E1010 14:47:06.233743 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:47:06 crc kubenswrapper[4788]: E1010 14:47:06.233951 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:47:07 crc kubenswrapper[4788]: I1010 14:47:07.234260 4788 scope.go:117] "RemoveContainer" containerID="a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be" Oct 10 14:47:07 crc kubenswrapper[4788]: I1010 14:47:07.947918 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/1.log" Oct 10 14:47:07 crc kubenswrapper[4788]: I1010 14:47:07.948231 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rzgfk" event={"ID":"39cfa1ec-4912-4550-9ead-7f6113db2221","Type":"ContainerStarted","Data":"e12ad039a71ffba781a71f8c01ee60626369df48ce03e2018de3b6a2dfc16c0f"} Oct 10 14:47:08 crc kubenswrapper[4788]: I1010 14:47:08.233384 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:08 crc kubenswrapper[4788]: I1010 14:47:08.233415 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:08 crc kubenswrapper[4788]: E1010 14:47:08.234931 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:47:08 crc kubenswrapper[4788]: I1010 14:47:08.234979 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:08 crc kubenswrapper[4788]: I1010 14:47:08.234959 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:08 crc kubenswrapper[4788]: E1010 14:47:08.235084 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:47:08 crc kubenswrapper[4788]: E1010 14:47:08.235229 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:47:08 crc kubenswrapper[4788]: E1010 14:47:08.235293 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:47:08 crc kubenswrapper[4788]: E1010 14:47:08.341276 4788 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 10 14:47:10 crc kubenswrapper[4788]: I1010 14:47:10.233120 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:10 crc kubenswrapper[4788]: E1010 14:47:10.233272 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:47:10 crc kubenswrapper[4788]: I1010 14:47:10.233352 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:10 crc kubenswrapper[4788]: I1010 14:47:10.233371 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:10 crc kubenswrapper[4788]: I1010 14:47:10.233409 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:10 crc kubenswrapper[4788]: E1010 14:47:10.233511 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:47:10 crc kubenswrapper[4788]: E1010 14:47:10.233662 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:47:10 crc kubenswrapper[4788]: E1010 14:47:10.233764 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:47:12 crc kubenswrapper[4788]: I1010 14:47:12.234275 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:12 crc kubenswrapper[4788]: I1010 14:47:12.234335 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:12 crc kubenswrapper[4788]: E1010 14:47:12.234429 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jqlwd" podUID="3d61e9aa-0688-436b-a8af-d6d8353c4350" Oct 10 14:47:12 crc kubenswrapper[4788]: I1010 14:47:12.234291 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:12 crc kubenswrapper[4788]: I1010 14:47:12.235321 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:12 crc kubenswrapper[4788]: E1010 14:47:12.235443 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 10 14:47:12 crc kubenswrapper[4788]: E1010 14:47:12.235586 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 10 14:47:12 crc kubenswrapper[4788]: E1010 14:47:12.237936 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.233749 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.233808 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.233749 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.234380 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.236037 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.236301 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.236422 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.237313 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.237899 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 10 14:47:14 crc kubenswrapper[4788]: I1010 14:47:14.239931 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.421485 4788 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.459341 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2n42n"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.460112 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.465065 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.465371 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.465461 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.466058 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.466186 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.467745 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.471379 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.471770 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.472555 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.472774 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m448m"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.473343 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.473811 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.476979 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.477465 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8txtm"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.477673 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.477960 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.479374 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.479503 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.480207 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.480694 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.481869 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.482319 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.482693 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.483073 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.483363 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.483958 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.484625 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.485044 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.485207 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.485343 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.485369 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.485437 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.485503 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.485653 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.485544 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.486000 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.486688 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.487197 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.488315 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.489103 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.489686 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.489884 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.490118 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.491072 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.493196 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.493539 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.495279 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9bw5d"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.495877 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.496018 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.496048 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.497085 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.497218 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.497655 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.497894 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-h28k6"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.498373 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-h28k6" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.498768 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.505959 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.506547 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.506774 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kdbvl"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.519041 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.519369 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.521352 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.521856 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.522542 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.523350 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.523702 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.524015 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.524820 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.525405 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.525438 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.526575 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.530088 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.531730 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.551009 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.551481 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5j9sp"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.551581 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.551799 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.551939 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.551941 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.552286 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.552369 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.551972 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.552463 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.552591 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.553612 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.553700 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.553733 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.553766 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.554616 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.555382 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cjsq9"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.555731 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.556036 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.556039 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.556474 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7xj4z"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.557166 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.557227 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.557970 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mgbhm"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.558462 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.558928 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qvvxg"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.559608 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.559899 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2n42n"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.560959 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.561446 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.571485 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.571816 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.571897 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-459pb"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.571832 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.572521 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.572652 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m448m"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.572750 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.572920 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.573082 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.573520 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9bw5d"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.573694 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.575757 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.580285 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.581270 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nrzg2"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.581533 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.581805 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.581899 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.581912 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.587558 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.588398 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.588567 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.588727 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.589443 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.589692 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.589732 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.589826 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.590012 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.590132 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.590216 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.590881 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7xj4z"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.590983 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.591514 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.591671 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.591858 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.591952 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.592021 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.592079 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.594705 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.594993 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.595599 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.603611 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.605174 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.605613 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.605899 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.605966 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.606097 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.606989 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.607281 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.622354 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624606 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-audit-dir\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624671 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-config\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624692 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-encryption-config\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624714 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5p2q\" (UniqueName: \"kubernetes.io/projected/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-kube-api-access-n5p2q\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624733 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-audit\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624754 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnj8c\" (UniqueName: \"kubernetes.io/projected/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-kube-api-access-qnj8c\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624774 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-config\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624790 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624822 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-serving-cert\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624842 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-serving-cert\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624862 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-config\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624878 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-client-ca\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624898 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-machine-approver-tls\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624915 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-image-import-ca\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.624960 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbjt9\" (UniqueName: \"kubernetes.io/projected/10e304aa-8568-440e-99cc-f9081412645e-kube-api-access-qbjt9\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625007 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxsxp\" (UniqueName: \"kubernetes.io/projected/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-kube-api-access-bxsxp\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625052 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48kt7\" (UniqueName: \"kubernetes.io/projected/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-kube-api-access-48kt7\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625073 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-images\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625116 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625158 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-config\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625184 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625204 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e304aa-8568-440e-99cc-f9081412645e-serving-cert\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625234 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-auth-proxy-config\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625252 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-etcd-serving-ca\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625271 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625354 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-node-pullsecrets\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.625393 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-etcd-client\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.627416 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.640275 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.640638 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.641518 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.641871 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.641915 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.642179 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.642534 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.642976 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.643248 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.644254 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.645445 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8vx6h"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.645978 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.646849 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.647268 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.648841 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.650354 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.651084 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.652792 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.653377 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-99fw8"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.653545 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.653780 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.654871 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wslmf"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.655777 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.656036 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.658338 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.658440 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.658906 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.659279 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.660006 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7zmwn"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.660779 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.661564 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.661886 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.662190 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.662824 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.663328 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.664274 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.664905 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.665748 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xfl9r"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.667477 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.667654 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nrzg2"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.669531 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.672682 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.674908 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cjsq9"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.677310 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.678728 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.682014 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-h28k6"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.682920 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.685124 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8txtm"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.686378 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.688259 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mgbhm"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.690266 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-459pb"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.692990 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qvvxg"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.694860 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5j9sp"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.696655 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.702997 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.706785 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wslmf"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.712222 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kdbvl"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.713951 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.715729 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8vx6h"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.716858 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-77m52"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.718127 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-77m52" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.719932 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6xm66"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.721998 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.723021 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7zmwn"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.723049 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.723062 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.723181 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.724197 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.725435 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xfl9r"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.725979 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5p2q\" (UniqueName: \"kubernetes.io/projected/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-kube-api-access-n5p2q\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726020 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-audit\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726049 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnj8c\" (UniqueName: \"kubernetes.io/projected/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-kube-api-access-qnj8c\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726072 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726095 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-serving-cert\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726117 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-config\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726162 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-serving-cert\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726191 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-config\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726211 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-client-ca\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726232 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-machine-approver-tls\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726254 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48kt7\" (UniqueName: \"kubernetes.io/projected/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-kube-api-access-48kt7\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726277 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-image-import-ca\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726296 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbjt9\" (UniqueName: \"kubernetes.io/projected/10e304aa-8568-440e-99cc-f9081412645e-kube-api-access-qbjt9\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726317 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxsxp\" (UniqueName: \"kubernetes.io/projected/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-kube-api-access-bxsxp\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726339 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-images\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726363 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726397 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-config\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726422 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726445 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e304aa-8568-440e-99cc-f9081412645e-serving-cert\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726471 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-auth-proxy-config\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726502 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-etcd-serving-ca\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726529 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726550 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-node-pullsecrets\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726589 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-etcd-client\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726614 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-audit-dir\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726645 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-config\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726676 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-encryption-config\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.726831 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-77m52"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.727331 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-node-pullsecrets\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.727487 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.727747 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-audit\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.728204 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-config\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.728525 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.728674 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.728699 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-image-import-ca\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.729729 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.730089 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.730295 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-auth-proxy-config\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.730348 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-client-ca\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.730386 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.730429 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-audit-dir\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.730771 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-config\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.730906 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-images\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.731115 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.731466 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-config\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.731706 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-etcd-serving-ca\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.732307 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6xm66"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.732494 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-serving-cert\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.732946 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-encryption-config\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.733031 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.733065 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-serving-cert\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.733355 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2g9kl"] Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.733584 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-config\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.733785 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-etcd-client\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.733972 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.734404 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e304aa-8568-440e-99cc-f9081412645e-serving-cert\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.735258 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-machine-approver-tls\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.742244 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.761796 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.781765 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.802089 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.822525 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.842767 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.862984 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.883431 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.903944 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.923645 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.948833 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.962122 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 10 14:47:17 crc kubenswrapper[4788]: I1010 14:47:17.982281 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.001497 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.023646 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.043013 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.063079 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.083644 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.116099 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.163806 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.183531 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.203747 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.223464 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.232920 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cedd44c5-a11e-48e9-b61f-04c087777fc9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jcksv\" (UID: \"cedd44c5-a11e-48e9-b61f-04c087777fc9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233064 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrms5\" (UniqueName: \"kubernetes.io/projected/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-kube-api-access-nrms5\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233126 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-serving-cert\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233244 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233288 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-config\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233330 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nt68\" (UniqueName: \"kubernetes.io/projected/51751bfc-73b2-4968-bbf7-a213e109b40a-kube-api-access-6nt68\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233369 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-etcd-client\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233409 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233513 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233587 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2821f67e-912c-48d7-8c95-47b29949afe9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233619 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-client-ca\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233655 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-certificates\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233670 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-trusted-ca\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233705 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb2p4\" (UniqueName: \"kubernetes.io/projected/7d2c7d67-2055-4e2b-9598-4438c3b702c1-kube-api-access-hb2p4\") pod \"downloads-7954f5f757-h28k6\" (UID: \"7d2c7d67-2055-4e2b-9598-4438c3b702c1\") " pod="openshift-console/downloads-7954f5f757-h28k6" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233781 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2821f67e-912c-48d7-8c95-47b29949afe9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233828 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51751bfc-73b2-4968-bbf7-a213e109b40a-serving-cert\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233846 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-encryption-config\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233871 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9nlv\" (UniqueName: \"kubernetes.io/projected/2821f67e-912c-48d7-8c95-47b29949afe9-kube-api-access-d9nlv\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.233889 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234006 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-bound-sa-token\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234032 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrqj5\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-kube-api-access-wrqj5\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234078 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-tls\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234101 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234164 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cddkn\" (UniqueName: \"kubernetes.io/projected/01c39c4c-1b9c-464e-aba0-434acf814e34-kube-api-access-cddkn\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234239 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpkph\" (UniqueName: \"kubernetes.io/projected/c389c203-6864-4423-bb2a-da5693e052dc-kube-api-access-gpkph\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234270 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234311 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v5hz\" (UniqueName: \"kubernetes.io/projected/cedd44c5-a11e-48e9-b61f-04c087777fc9-kube-api-access-2v5hz\") pod \"cluster-samples-operator-665b6dd947-jcksv\" (UID: \"cedd44c5-a11e-48e9-b61f-04c087777fc9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234337 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c39c4c-1b9c-464e-aba0-434acf814e34-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234360 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-audit-policies\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234382 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c39c4c-1b9c-464e-aba0-434acf814e34-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234416 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234481 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c389c203-6864-4423-bb2a-da5693e052dc-audit-dir\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.234500 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.235068 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:18.735037532 +0000 UTC m=+141.184753300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.242546 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.262391 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.283202 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.303717 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.323490 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335030 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.335225 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:18.835201557 +0000 UTC m=+141.284917105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335364 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ce4a5b7-557b-4b26-924c-2e246030b180-service-ca-bundle\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335397 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f007d78-6f96-4523-8ae6-7d03955f0743-config\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335444 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb2p4\" (UniqueName: \"kubernetes.io/projected/7d2c7d67-2055-4e2b-9598-4438c3b702c1-kube-api-access-hb2p4\") pod \"downloads-7954f5f757-h28k6\" (UID: \"7d2c7d67-2055-4e2b-9598-4438c3b702c1\") " pod="openshift-console/downloads-7954f5f757-h28k6" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335467 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47d58cb0-69cb-485b-b1e1-c857093a4c3c-metrics-tls\") pod \"dns-operator-744455d44c-7xj4z\" (UID: \"47d58cb0-69cb-485b-b1e1-c857093a4c3c\") " pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335504 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdzc8\" (UniqueName: \"kubernetes.io/projected/47d58cb0-69cb-485b-b1e1-c857093a4c3c-kube-api-access-mdzc8\") pod \"dns-operator-744455d44c-7xj4z\" (UID: \"47d58cb0-69cb-485b-b1e1-c857093a4c3c\") " pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335527 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-encryption-config\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335545 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501c7fd2-e29f-48d8-9cf5-655af1e87402-config-volume\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335584 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335607 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-csi-data-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335676 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-etcd-service-ca\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335711 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9nlv\" (UniqueName: \"kubernetes.io/projected/2821f67e-912c-48d7-8c95-47b29949afe9-kube-api-access-d9nlv\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335749 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335767 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f0008884-e2e1-4229-ae6d-86bd85c7c301-srv-cert\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335784 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f0008884-e2e1-4229-ae6d-86bd85c7c301-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335900 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v225m\" (UniqueName: \"kubernetes.io/projected/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-kube-api-access-v225m\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335931 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b667bac-e257-4741-b4e7-38857403b391-config\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.335950 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-default-certificate\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336004 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-bound-sa-token\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336077 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrqj5\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-kube-api-access-wrqj5\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336100 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e6875754-4b83-4a13-b8e6-329300988b54-profile-collector-cert\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336116 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05e27a21-9c12-471b-b959-0a8a2046a371-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336196 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-tls\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336217 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336259 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dngzx\" (UniqueName: \"kubernetes.io/projected/360b557a-fc75-4704-8cb2-fb9285d18566-kube-api-access-dngzx\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336328 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cddkn\" (UniqueName: \"kubernetes.io/projected/01c39c4c-1b9c-464e-aba0-434acf814e34-kube-api-access-cddkn\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336348 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpkph\" (UniqueName: \"kubernetes.io/projected/c389c203-6864-4423-bb2a-da5693e052dc-kube-api-access-gpkph\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336366 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336440 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336568 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v5hz\" (UniqueName: \"kubernetes.io/projected/cedd44c5-a11e-48e9-b61f-04c087777fc9-kube-api-access-2v5hz\") pod \"cluster-samples-operator-665b6dd947-jcksv\" (UID: \"cedd44c5-a11e-48e9-b61f-04c087777fc9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336688 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e6875754-4b83-4a13-b8e6-329300988b54-srv-cert\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336709 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c39c4c-1b9c-464e-aba0-434acf814e34-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336806 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c54b345b-4399-418f-9bf2-a4c81e8c558b-cert\") pod \"ingress-canary-77m52\" (UID: \"c54b345b-4399-418f-9bf2-a4c81e8c558b\") " pod="openshift-ingress-canary/ingress-canary-77m52" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336825 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19cfebf6-d761-4be2-a146-082c2a71deba-trusted-ca\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336843 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad02b5d-b3d9-4005-90d4-8e824d28a95d-serving-cert\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336868 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf4nc\" (UniqueName: \"kubernetes.io/projected/1ce4a5b7-557b-4b26-924c-2e246030b180-kube-api-access-hf4nc\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.336965 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.337111 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.337133 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/501c7fd2-e29f-48d8-9cf5-655af1e87402-metrics-tls\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.337184 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-serving-cert\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.337204 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr9vv\" (UniqueName: \"kubernetes.io/projected/e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7-kube-api-access-nr9vv\") pod \"migrator-59844c95c7-2bt9s\" (UID: \"e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.337357 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b667bac-e257-4741-b4e7-38857403b391-serving-cert\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.337389 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c39c4c-1b9c-464e-aba0-434acf814e34-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.337444 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-oauth-config\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.337616 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-policies\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338177 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338301 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338361 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338486 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv7vk\" (UniqueName: \"kubernetes.io/projected/d8826b11-1579-4653-953e-0895bc1d16b5-kube-api-access-kv7vk\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338490 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-encryption-config\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338541 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338571 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztkhv\" (UniqueName: \"kubernetes.io/projected/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-kube-api-access-ztkhv\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338607 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338637 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-stats-auth\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338732 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338798 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-dir\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338840 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338868 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgkbb\" (UniqueName: \"kubernetes.io/projected/c54b345b-4399-418f-9bf2-a4c81e8c558b-kube-api-access-hgkbb\") pod \"ingress-canary-77m52\" (UID: \"c54b345b-4399-418f-9bf2-a4c81e8c558b\") " pod="openshift-ingress-canary/ingress-canary-77m52" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338885 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gnrr\" (UniqueName: \"kubernetes.io/projected/ffc68803-aa9e-4151-abc6-6d6a3eef7a8b-kube-api-access-8gnrr\") pod \"multus-admission-controller-857f4d67dd-8vx6h\" (UID: \"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338923 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlss6\" (UniqueName: \"kubernetes.io/projected/2686886b-00d1-4e24-8a84-f634c766fcca-kube-api-access-wlss6\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338943 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt79c\" (UniqueName: \"kubernetes.io/projected/5b667bac-e257-4741-b4e7-38857403b391-kube-api-access-wt79c\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.338960 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfvb7\" (UniqueName: \"kubernetes.io/projected/e6875754-4b83-4a13-b8e6-329300988b54-kube-api-access-nfvb7\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339015 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-trusted-ca\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339032 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f7897a5f-602b-4d6d-b266-4a80080d8671-tmpfs\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339050 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339092 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6gtc\" (UniqueName: \"kubernetes.io/projected/bb4a9174-7055-4a98-a3dc-790713348443-kube-api-access-v6gtc\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339114 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-config\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339135 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339178 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2821f67e-912c-48d7-8c95-47b29949afe9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339199 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wrlf\" (UniqueName: \"kubernetes.io/projected/6b2c9335-c98e-4808-b9cf-ecac89cd1a99-kube-api-access-9wrlf\") pod \"control-plane-machine-set-operator-78cbb6b69f-cgbfx\" (UID: \"6b2c9335-c98e-4808-b9cf-ecac89cd1a99\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339216 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f007d78-6f96-4523-8ae6-7d03955f0743-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339255 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-serving-cert\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339273 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-metrics-certs\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339321 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-certificates\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339407 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-trusted-ca\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339470 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339512 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339525 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05e27a21-9c12-471b-b959-0a8a2046a371-proxy-tls\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339570 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f007d78-6f96-4523-8ae6-7d03955f0743-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339612 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f7897a5f-602b-4d6d-b266-4a80080d8671-apiservice-cert\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339652 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-images\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339866 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2821f67e-912c-48d7-8c95-47b29949afe9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.339890 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51751bfc-73b2-4968-bbf7-a213e109b40a-serving-cert\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340052 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpwwz\" (UniqueName: \"kubernetes.io/projected/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-kube-api-access-jpwwz\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340596 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khxqz\" (UniqueName: \"kubernetes.io/projected/f7897a5f-602b-4d6d-b266-4a80080d8671-kube-api-access-khxqz\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340616 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-certs\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340170 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340534 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2821f67e-912c-48d7-8c95-47b29949afe9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340855 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340877 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b2c9335-c98e-4808-b9cf-ecac89cd1a99-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cgbfx\" (UID: \"6b2c9335-c98e-4808-b9cf-ecac89cd1a99\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340915 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f28vg\" (UniqueName: \"kubernetes.io/projected/19cfebf6-d761-4be2-a146-082c2a71deba-kube-api-access-f28vg\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.340989 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341010 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9x89\" (UniqueName: \"kubernetes.io/projected/aad02b5d-b3d9-4005-90d4-8e824d28a95d-kube-api-access-x9x89\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341065 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341084 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9pz2\" (UniqueName: \"kubernetes.io/projected/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-kube-api-access-w9pz2\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341101 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4n2l\" (UniqueName: \"kubernetes.io/projected/c3060c66-e983-43ae-b021-0a5c9b6dc99b-kube-api-access-r4n2l\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341219 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-audit-policies\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341277 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-service-ca\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341311 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c39c4c-1b9c-464e-aba0-434acf814e34-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341341 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/19cfebf6-d761-4be2-a146-082c2a71deba-metrics-tls\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341368 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-service-ca-bundle\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341406 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341438 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7199dcec-2a4b-418b-ba9b-d04810aa6e70-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341595 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c389c203-6864-4423-bb2a-da5693e052dc-audit-dir\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341638 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c389c203-6864-4423-bb2a-da5693e052dc-audit-dir\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341685 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f7897a5f-602b-4d6d-b266-4a80080d8671-webhook-cert\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341710 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-socket-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341734 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19cfebf6-d761-4be2-a146-082c2a71deba-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.341763 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:18.841749914 +0000 UTC m=+141.291465462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341785 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341812 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-proxy-tls\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341835 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1391a9-17fa-48f0-9145-6356860930f8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341851 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-config\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341865 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7199dcec-2a4b-418b-ba9b-d04810aa6e70-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341880 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq9rs\" (UniqueName: \"kubernetes.io/projected/05e27a21-9c12-471b-b959-0a8a2046a371-kube-api-access-dq9rs\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341896 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-signing-key\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341917 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgsks\" (UniqueName: \"kubernetes.io/projected/ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7-kube-api-access-mgsks\") pod \"package-server-manager-789f6589d5-2cft7\" (UID: \"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341943 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341960 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-oauth-serving-cert\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341982 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb4a9174-7055-4a98-a3dc-790713348443-etcd-client\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341949 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-trusted-ca\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.341997 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-signing-cabundle\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342096 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cedd44c5-a11e-48e9-b61f-04c087777fc9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jcksv\" (UID: \"cedd44c5-a11e-48e9-b61f-04c087777fc9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342255 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjhpt\" (UniqueName: \"kubernetes.io/projected/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-kube-api-access-gjhpt\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342314 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ffc68803-aa9e-4151-abc6-6d6a3eef7a8b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8vx6h\" (UID: \"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342337 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2821f67e-912c-48d7-8c95-47b29949afe9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342371 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lglr9\" (UniqueName: \"kubernetes.io/projected/7199dcec-2a4b-418b-ba9b-d04810aa6e70-kube-api-access-lglr9\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342478 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee1391a9-17fa-48f0-9145-6356860930f8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342547 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-plugins-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342599 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrms5\" (UniqueName: \"kubernetes.io/projected/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-kube-api-access-nrms5\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342654 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5lpl\" (UniqueName: \"kubernetes.io/projected/f0008884-e2e1-4229-ae6d-86bd85c7c301-kube-api-access-c5lpl\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.342941 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c389c203-6864-4423-bb2a-da5693e052dc-audit-policies\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343037 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-certificates\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343250 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343534 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-trusted-ca-bundle\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343583 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-etcd-ca\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343605 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-config-volume\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343625 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343669 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-node-bootstrap-token\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343691 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-serving-cert\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343709 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-config\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343759 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-console-config\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343777 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-registration-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343840 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343889 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb4a9174-7055-4a98-a3dc-790713348443-serving-cert\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343931 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-config\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343948 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nt68\" (UniqueName: \"kubernetes.io/projected/51751bfc-73b2-4968-bbf7-a213e109b40a-kube-api-access-6nt68\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343966 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxrtw\" (UniqueName: \"kubernetes.io/projected/501c7fd2-e29f-48d8-9cf5-655af1e87402-kube-api-access-cxrtw\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.343983 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2cft7\" (UID: \"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.344005 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-etcd-client\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.344021 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-mountpoint-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.344039 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee1391a9-17fa-48f0-9145-6356860930f8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.344062 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-client-ca\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.344081 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.344098 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-config\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.344125 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-secret-volume\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.344156 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.345765 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-client-ca\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.346077 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-config\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.346102 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-serving-cert\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.346266 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51751bfc-73b2-4968-bbf7-a213e109b40a-serving-cert\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.347305 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.347381 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cedd44c5-a11e-48e9-b61f-04c087777fc9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-jcksv\" (UID: \"cedd44c5-a11e-48e9-b61f-04c087777fc9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.350716 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.352672 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c39c4c-1b9c-464e-aba0-434acf814e34-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.353004 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-tls\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.355852 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c389c203-6864-4423-bb2a-da5693e052dc-etcd-client\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.363933 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.383278 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.402659 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.422730 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.442891 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445445 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.445557 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:18.945529505 +0000 UTC m=+141.395245053 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445651 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxrtw\" (UniqueName: \"kubernetes.io/projected/501c7fd2-e29f-48d8-9cf5-655af1e87402-kube-api-access-cxrtw\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445680 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb4a9174-7055-4a98-a3dc-790713348443-serving-cert\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445701 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-mountpoint-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445716 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee1391a9-17fa-48f0-9145-6356860930f8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445732 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2cft7\" (UID: \"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445753 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445769 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-config\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445786 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445805 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-secret-volume\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445819 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f007d78-6f96-4523-8ae6-7d03955f0743-config\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445838 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ce4a5b7-557b-4b26-924c-2e246030b180-service-ca-bundle\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445860 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47d58cb0-69cb-485b-b1e1-c857093a4c3c-metrics-tls\") pod \"dns-operator-744455d44c-7xj4z\" (UID: \"47d58cb0-69cb-485b-b1e1-c857093a4c3c\") " pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445876 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdzc8\" (UniqueName: \"kubernetes.io/projected/47d58cb0-69cb-485b-b1e1-c857093a4c3c-kube-api-access-mdzc8\") pod \"dns-operator-744455d44c-7xj4z\" (UID: \"47d58cb0-69cb-485b-b1e1-c857093a4c3c\") " pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445897 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501c7fd2-e29f-48d8-9cf5-655af1e87402-config-volume\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445915 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445940 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f0008884-e2e1-4229-ae6d-86bd85c7c301-srv-cert\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445954 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-csi-data-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445972 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-etcd-service-ca\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.445997 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f0008884-e2e1-4229-ae6d-86bd85c7c301-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446016 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v225m\" (UniqueName: \"kubernetes.io/projected/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-kube-api-access-v225m\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446026 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-mountpoint-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446047 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e6875754-4b83-4a13-b8e6-329300988b54-profile-collector-cert\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446064 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b667bac-e257-4741-b4e7-38857403b391-config\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446080 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-default-certificate\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446099 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dngzx\" (UniqueName: \"kubernetes.io/projected/360b557a-fc75-4704-8cb2-fb9285d18566-kube-api-access-dngzx\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446117 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05e27a21-9c12-471b-b959-0a8a2046a371-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446162 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446187 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e6875754-4b83-4a13-b8e6-329300988b54-srv-cert\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446203 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c54b345b-4399-418f-9bf2-a4c81e8c558b-cert\") pod \"ingress-canary-77m52\" (UID: \"c54b345b-4399-418f-9bf2-a4c81e8c558b\") " pod="openshift-ingress-canary/ingress-canary-77m52" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446219 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad02b5d-b3d9-4005-90d4-8e824d28a95d-serving-cert\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446235 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf4nc\" (UniqueName: \"kubernetes.io/projected/1ce4a5b7-557b-4b26-924c-2e246030b180-kube-api-access-hf4nc\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446253 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19cfebf6-d761-4be2-a146-082c2a71deba-trusted-ca\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446272 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/501c7fd2-e29f-48d8-9cf5-655af1e87402-metrics-tls\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446288 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-serving-cert\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446305 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr9vv\" (UniqueName: \"kubernetes.io/projected/e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7-kube-api-access-nr9vv\") pod \"migrator-59844c95c7-2bt9s\" (UID: \"e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446322 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b667bac-e257-4741-b4e7-38857403b391-serving-cert\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446339 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-oauth-config\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446354 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-policies\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446373 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446389 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446405 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446419 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv7vk\" (UniqueName: \"kubernetes.io/projected/d8826b11-1579-4653-953e-0895bc1d16b5-kube-api-access-kv7vk\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446434 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446453 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztkhv\" (UniqueName: \"kubernetes.io/projected/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-kube-api-access-ztkhv\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446469 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446501 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-stats-auth\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446519 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446520 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f007d78-6f96-4523-8ae6-7d03955f0743-config\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446551 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-dir\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446574 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446640 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgkbb\" (UniqueName: \"kubernetes.io/projected/c54b345b-4399-418f-9bf2-a4c81e8c558b-kube-api-access-hgkbb\") pod \"ingress-canary-77m52\" (UID: \"c54b345b-4399-418f-9bf2-a4c81e8c558b\") " pod="openshift-ingress-canary/ingress-canary-77m52" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446662 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt79c\" (UniqueName: \"kubernetes.io/projected/5b667bac-e257-4741-b4e7-38857403b391-kube-api-access-wt79c\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446681 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gnrr\" (UniqueName: \"kubernetes.io/projected/ffc68803-aa9e-4151-abc6-6d6a3eef7a8b-kube-api-access-8gnrr\") pod \"multus-admission-controller-857f4d67dd-8vx6h\" (UID: \"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446700 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlss6\" (UniqueName: \"kubernetes.io/projected/2686886b-00d1-4e24-8a84-f634c766fcca-kube-api-access-wlss6\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446718 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-trusted-ca\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446736 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f7897a5f-602b-4d6d-b266-4a80080d8671-tmpfs\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446754 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfvb7\" (UniqueName: \"kubernetes.io/projected/e6875754-4b83-4a13-b8e6-329300988b54-kube-api-access-nfvb7\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446774 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-config\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446790 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6gtc\" (UniqueName: \"kubernetes.io/projected/bb4a9174-7055-4a98-a3dc-790713348443-kube-api-access-v6gtc\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446808 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wrlf\" (UniqueName: \"kubernetes.io/projected/6b2c9335-c98e-4808-b9cf-ecac89cd1a99-kube-api-access-9wrlf\") pod \"control-plane-machine-set-operator-78cbb6b69f-cgbfx\" (UID: \"6b2c9335-c98e-4808-b9cf-ecac89cd1a99\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446828 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f007d78-6f96-4523-8ae6-7d03955f0743-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446845 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-serving-cert\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446860 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-metrics-certs\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446876 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446896 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05e27a21-9c12-471b-b959-0a8a2046a371-proxy-tls\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446915 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f007d78-6f96-4523-8ae6-7d03955f0743-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446932 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-images\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446949 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f7897a5f-602b-4d6d-b266-4a80080d8671-apiservice-cert\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446965 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpwwz\" (UniqueName: \"kubernetes.io/projected/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-kube-api-access-jpwwz\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.446988 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khxqz\" (UniqueName: \"kubernetes.io/projected/f7897a5f-602b-4d6d-b266-4a80080d8671-kube-api-access-khxqz\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447005 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-certs\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447021 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b2c9335-c98e-4808-b9cf-ecac89cd1a99-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cgbfx\" (UID: \"6b2c9335-c98e-4808-b9cf-ecac89cd1a99\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447046 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447063 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9x89\" (UniqueName: \"kubernetes.io/projected/aad02b5d-b3d9-4005-90d4-8e824d28a95d-kube-api-access-x9x89\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447080 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f28vg\" (UniqueName: \"kubernetes.io/projected/19cfebf6-d761-4be2-a146-082c2a71deba-kube-api-access-f28vg\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447095 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447118 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9pz2\" (UniqueName: \"kubernetes.io/projected/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-kube-api-access-w9pz2\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447157 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4n2l\" (UniqueName: \"kubernetes.io/projected/c3060c66-e983-43ae-b021-0a5c9b6dc99b-kube-api-access-r4n2l\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447177 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-service-ca\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447192 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-service-ca-bundle\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447208 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/19cfebf6-d761-4be2-a146-082c2a71deba-metrics-tls\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447221 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-config\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447229 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447253 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7199dcec-2a4b-418b-ba9b-d04810aa6e70-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447279 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19cfebf6-d761-4be2-a146-082c2a71deba-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447296 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447311 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-proxy-tls\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447327 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f7897a5f-602b-4d6d-b266-4a80080d8671-webhook-cert\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447343 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-socket-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447358 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1391a9-17fa-48f0-9145-6356860930f8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447377 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7199dcec-2a4b-418b-ba9b-d04810aa6e70-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447393 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq9rs\" (UniqueName: \"kubernetes.io/projected/05e27a21-9c12-471b-b959-0a8a2046a371-kube-api-access-dq9rs\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447410 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-signing-key\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447428 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgsks\" (UniqueName: \"kubernetes.io/projected/ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7-kube-api-access-mgsks\") pod \"package-server-manager-789f6589d5-2cft7\" (UID: \"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447445 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-config\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447461 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447479 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-oauth-serving-cert\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447495 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb4a9174-7055-4a98-a3dc-790713348443-etcd-client\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447519 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjhpt\" (UniqueName: \"kubernetes.io/projected/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-kube-api-access-gjhpt\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447537 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ffc68803-aa9e-4151-abc6-6d6a3eef7a8b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8vx6h\" (UID: \"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447553 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lglr9\" (UniqueName: \"kubernetes.io/projected/7199dcec-2a4b-418b-ba9b-d04810aa6e70-kube-api-access-lglr9\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447570 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-signing-cabundle\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447585 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee1391a9-17fa-48f0-9145-6356860930f8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447601 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-plugins-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447628 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5lpl\" (UniqueName: \"kubernetes.io/projected/f0008884-e2e1-4229-ae6d-86bd85c7c301-kube-api-access-c5lpl\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447644 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-trusted-ca-bundle\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447663 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-etcd-ca\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447681 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-config-volume\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.447901 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-dir\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448081 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448116 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-node-bootstrap-token\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448157 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-config\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448179 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-console-config\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448202 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-registration-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448271 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/05e27a21-9c12-471b-b959-0a8a2046a371-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448395 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448433 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-csi-data-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448457 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-registration-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.448777 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-policies\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.449413 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb4a9174-7055-4a98-a3dc-790713348443-serving-cert\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.449803 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.449821 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19cfebf6-d761-4be2-a146-082c2a71deba-trusted-ca\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.449837 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.449987 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-trusted-ca\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.450560 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f7897a5f-602b-4d6d-b266-4a80080d8671-tmpfs\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.451327 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.451423 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-socket-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.451420 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.451522 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.451548 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/47d58cb0-69cb-485b-b1e1-c857093a4c3c-metrics-tls\") pod \"dns-operator-744455d44c-7xj4z\" (UID: \"47d58cb0-69cb-485b-b1e1-c857093a4c3c\") " pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.451957 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.451963 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-config\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.451978 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:18.951965419 +0000 UTC m=+141.401680957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.452041 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/360b557a-fc75-4704-8cb2-fb9285d18566-plugins-dir\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.452234 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-serving-cert\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.452695 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.452998 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-config\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.453757 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee1391a9-17fa-48f0-9145-6356860930f8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.454126 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aad02b5d-b3d9-4005-90d4-8e824d28a95d-serving-cert\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.454197 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-etcd-ca\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.454357 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-images\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.454422 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.454572 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-proxy-tls\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.454709 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-oauth-config\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.454861 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.455367 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.455433 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-serving-cert\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.455744 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-trusted-ca-bundle\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.455858 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.456388 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-console-config\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.456497 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-oauth-serving-cert\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.456707 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-service-ca\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.456972 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.457299 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.457455 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.457567 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bb4a9174-7055-4a98-a3dc-790713348443-etcd-service-ca\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.458390 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/19cfebf6-d761-4be2-a146-082c2a71deba-metrics-tls\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.458736 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad02b5d-b3d9-4005-90d4-8e824d28a95d-service-ca-bundle\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.458814 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7199dcec-2a4b-418b-ba9b-d04810aa6e70-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.458919 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee1391a9-17fa-48f0-9145-6356860930f8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.460710 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb4a9174-7055-4a98-a3dc-790713348443-etcd-client\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.462651 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.463051 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f007d78-6f96-4523-8ae6-7d03955f0743-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.483260 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.491412 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7199dcec-2a4b-418b-ba9b-d04810aa6e70-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.502794 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.523655 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.543462 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.548659 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.548855 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.548904 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.048877905 +0000 UTC m=+141.498593443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.549595 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.550039 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.050025702 +0000 UTC m=+141.499741460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.562260 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.573370 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-config\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.583610 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.603131 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.608583 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ffc68803-aa9e-4151-abc6-6d6a3eef7a8b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8vx6h\" (UID: \"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.622324 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.643433 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.648624 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b2c9335-c98e-4808-b9cf-ecac89cd1a99-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-cgbfx\" (UID: \"6b2c9335-c98e-4808-b9cf-ecac89cd1a99\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.650573 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.650741 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.150714329 +0000 UTC m=+141.600429897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.650996 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.651586 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.15156897 +0000 UTC m=+141.601284508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.661292 4788 request.go:700] Waited for 1.008359621s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-sa-dockercfg-5xfcg&limit=500&resourceVersion=0 Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.663027 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.682854 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.703414 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.722327 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.730761 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/05e27a21-9c12-471b-b959-0a8a2046a371-proxy-tls\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.743259 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.752581 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.752750 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.252730007 +0000 UTC m=+141.702445555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.752844 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.753249 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.25324177 +0000 UTC m=+141.702957318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.762264 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.767258 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-metrics-certs\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.782395 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.787652 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1ce4a5b7-557b-4b26-924c-2e246030b180-service-ca-bundle\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.811855 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.821711 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-default-certificate\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.822502 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.832211 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1ce4a5b7-557b-4b26-924c-2e246030b180-stats-auth\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.843362 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.854498 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.854697 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.354668004 +0000 UTC m=+141.804383742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.854875 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.855238 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.355224317 +0000 UTC m=+141.804939865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.862502 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.882085 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.902316 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.922846 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.930551 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5b667bac-e257-4741-b4e7-38857403b391-serving-cert\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.943187 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.949416 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5b667bac-e257-4741-b4e7-38857403b391-config\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.956748 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.956895 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.456860926 +0000 UTC m=+141.906576474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.957548 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:18 crc kubenswrapper[4788]: E1010 14:47:18.958017 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.457996664 +0000 UTC m=+141.907712222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.963624 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 10 14:47:18 crc kubenswrapper[4788]: I1010 14:47:18.982560 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.002363 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.023463 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.043734 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.052047 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f0008884-e2e1-4229-ae6d-86bd85c7c301-srv-cert\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.059307 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.059556 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.55952553 +0000 UTC m=+142.009241098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.060200 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.060811 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.560801611 +0000 UTC m=+142.010517179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.063358 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.069593 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f0008884-e2e1-4229-ae6d-86bd85c7c301-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.070785 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e6875754-4b83-4a13-b8e6-329300988b54-profile-collector-cert\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.071262 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-secret-volume\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.083598 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.102390 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.104918 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-config-volume\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.123067 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.143397 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.148504 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-signing-key\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.162309 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.162911 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.662886061 +0000 UTC m=+142.112601609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.163276 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.164055 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.664023548 +0000 UTC m=+142.113739136 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.164989 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.183698 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.202646 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.214623 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-signing-cabundle\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.223873 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.239421 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.244217 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.254632 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e6875754-4b83-4a13-b8e6-329300988b54-srv-cert\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.263627 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.264494 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.264539 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.76451437 +0000 UTC m=+142.214229918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.265837 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.266183 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.76617375 +0000 UTC m=+142.215889298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.269459 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f7897a5f-602b-4d6d-b266-4a80080d8671-apiservice-cert\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.277217 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f7897a5f-602b-4d6d-b266-4a80080d8671-webhook-cert\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.283809 4788 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.303601 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.324356 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.345118 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.352669 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2cft7\" (UID: \"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.363632 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.367273 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.367501 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.867462491 +0000 UTC m=+142.317178079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.367980 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.368919 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.868892245 +0000 UTC m=+142.318607823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.376321 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c54b345b-4399-418f-9bf2-a4c81e8c558b-cert\") pod \"ingress-canary-77m52\" (UID: \"c54b345b-4399-418f-9bf2-a4c81e8c558b\") " pod="openshift-ingress-canary/ingress-canary-77m52" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.383064 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.404275 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.424097 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.443884 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.448240 4788 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.448347 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/501c7fd2-e29f-48d8-9cf5-655af1e87402-config-volume podName:501c7fd2-e29f-48d8-9cf5-655af1e87402 nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.948326212 +0000 UTC m=+142.398041760 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/501c7fd2-e29f-48d8-9cf5-655af1e87402-config-volume") pod "dns-default-6xm66" (UID: "501c7fd2-e29f-48d8-9cf5-655af1e87402") : failed to sync configmap cache: timed out waiting for the condition Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.448579 4788 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.448617 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/501c7fd2-e29f-48d8-9cf5-655af1e87402-metrics-tls podName:501c7fd2-e29f-48d8-9cf5-655af1e87402 nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.948609908 +0000 UTC m=+142.398325456 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/501c7fd2-e29f-48d8-9cf5-655af1e87402-metrics-tls") pod "dns-default-6xm66" (UID: "501c7fd2-e29f-48d8-9cf5-655af1e87402") : failed to sync secret cache: timed out waiting for the condition Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.453965 4788 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.454094 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-node-bootstrap-token podName:c3060c66-e983-43ae-b021-0a5c9b6dc99b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.954064659 +0000 UTC m=+142.403780237 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-node-bootstrap-token") pod "machine-config-server-2g9kl" (UID: "c3060c66-e983-43ae-b021-0a5c9b6dc99b") : failed to sync secret cache: timed out waiting for the condition Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.455767 4788 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.455829 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-certs podName:c3060c66-e983-43ae-b021-0a5c9b6dc99b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.955814991 +0000 UTC m=+142.405530539 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-certs") pod "machine-config-server-2g9kl" (UID: "c3060c66-e983-43ae-b021-0a5c9b6dc99b") : failed to sync secret cache: timed out waiting for the condition Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.464072 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.470194 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.470417 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.970377561 +0000 UTC m=+142.420093109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.470738 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.471222 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:19.971211331 +0000 UTC m=+142.420927079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.484397 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.530447 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbjt9\" (UniqueName: \"kubernetes.io/projected/10e304aa-8568-440e-99cc-f9081412645e-kube-api-access-qbjt9\") pod \"controller-manager-879f6c89f-m448m\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.541762 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5p2q\" (UniqueName: \"kubernetes.io/projected/fcf134f5-5c52-447f-b5b7-d3fbe2aef728-kube-api-access-n5p2q\") pod \"apiserver-76f77b778f-2n42n\" (UID: \"fcf134f5-5c52-447f-b5b7-d3fbe2aef728\") " pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.567415 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48kt7\" (UniqueName: \"kubernetes.io/projected/cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0-kube-api-access-48kt7\") pod \"openshift-config-operator-7777fb866f-8txtm\" (UID: \"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.573070 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.573283 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.07324648 +0000 UTC m=+142.522962068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.573849 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.574275 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.074259624 +0000 UTC m=+142.523975172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.582808 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.591513 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnj8c\" (UniqueName: \"kubernetes.io/projected/a46b6887-dafb-4cd1-a4f6-2a9602b84b89-kube-api-access-qnj8c\") pod \"machine-api-operator-5694c8668f-9bw5d\" (UID: \"a46b6887-dafb-4cd1-a4f6-2a9602b84b89\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.601885 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxsxp\" (UniqueName: \"kubernetes.io/projected/99fc4a35-5cf5-41a6-82a9-0814ac116d7d-kube-api-access-bxsxp\") pod \"machine-approver-56656f9798-vwlxm\" (UID: \"99fc4a35-5cf5-41a6-82a9-0814ac116d7d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.603083 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.623603 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.624683 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.643840 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.661516 4788 request.go:700] Waited for 1.524433377s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-5j9sp Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.661679 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.662081 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.675410 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.676564 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.176520328 +0000 UTC m=+142.626235896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.676883 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.677480 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.177449991 +0000 UTC m=+142.627165569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: W1010 14:47:19.687664 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99fc4a35_5cf5_41a6_82a9_0814ac116d7d.slice/crio-42c9538ef4be62c4a365b2cc5d1f09ef75a06fb5b3e96375bc663696b347e168 WatchSource:0}: Error finding container 42c9538ef4be62c4a365b2cc5d1f09ef75a06fb5b3e96375bc663696b347e168: Status 404 returned error can't find the container with id 42c9538ef4be62c4a365b2cc5d1f09ef75a06fb5b3e96375bc663696b347e168 Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.701900 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb2p4\" (UniqueName: \"kubernetes.io/projected/7d2c7d67-2055-4e2b-9598-4438c3b702c1-kube-api-access-hb2p4\") pod \"downloads-7954f5f757-h28k6\" (UID: \"7d2c7d67-2055-4e2b-9598-4438c3b702c1\") " pod="openshift-console/downloads-7954f5f757-h28k6" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.744992 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-bound-sa-token\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.746470 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9nlv\" (UniqueName: \"kubernetes.io/projected/2821f67e-912c-48d7-8c95-47b29949afe9-kube-api-access-d9nlv\") pod \"openshift-apiserver-operator-796bbdcf4f-hvvxp\" (UID: \"2821f67e-912c-48d7-8c95-47b29949afe9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.758986 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.768394 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cddkn\" (UniqueName: \"kubernetes.io/projected/01c39c4c-1b9c-464e-aba0-434acf814e34-kube-api-access-cddkn\") pod \"openshift-controller-manager-operator-756b6f6bc6-8sb2v\" (UID: \"01c39c4c-1b9c-464e-aba0-434acf814e34\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.782786 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.785816 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-h28k6" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.786014 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.285991676 +0000 UTC m=+142.735707224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.786750 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.787197 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.287171844 +0000 UTC m=+142.736887392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.788641 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpkph\" (UniqueName: \"kubernetes.io/projected/c389c203-6864-4423-bb2a-da5693e052dc-kube-api-access-gpkph\") pod \"apiserver-7bbb656c7d-n2bnk\" (UID: \"c389c203-6864-4423-bb2a-da5693e052dc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.815863 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v5hz\" (UniqueName: \"kubernetes.io/projected/cedd44c5-a11e-48e9-b61f-04c087777fc9-kube-api-access-2v5hz\") pod \"cluster-samples-operator-665b6dd947-jcksv\" (UID: \"cedd44c5-a11e-48e9-b61f-04c087777fc9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.819357 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrqj5\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-kube-api-access-wrqj5\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.841764 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.858090 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrms5\" (UniqueName: \"kubernetes.io/projected/2e6e895e-b3ff-4a54-9147-0e5e962bbdc4-kube-api-access-nrms5\") pod \"cluster-image-registry-operator-dc59b4c8b-h7d2s\" (UID: \"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.876082 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2n42n"] Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.883106 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nt68\" (UniqueName: \"kubernetes.io/projected/51751bfc-73b2-4968-bbf7-a213e109b40a-kube-api-access-6nt68\") pod \"route-controller-manager-6576b87f9c-rnqfq\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.890244 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.890563 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.390529595 +0000 UTC m=+142.840245283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.890937 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.891456 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.391447386 +0000 UTC m=+142.841162934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.905733 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxrtw\" (UniqueName: \"kubernetes.io/projected/501c7fd2-e29f-48d8-9cf5-655af1e87402-kube-api-access-cxrtw\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.925532 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee1391a9-17fa-48f0-9145-6356860930f8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k74dd\" (UID: \"ee1391a9-17fa-48f0-9145-6356860930f8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.946819 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdzc8\" (UniqueName: \"kubernetes.io/projected/47d58cb0-69cb-485b-b1e1-c857093a4c3c-kube-api-access-mdzc8\") pod \"dns-operator-744455d44c-7xj4z\" (UID: \"47d58cb0-69cb-485b-b1e1-c857093a4c3c\") " pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.966778 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr9vv\" (UniqueName: \"kubernetes.io/projected/e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7-kube-api-access-nr9vv\") pod \"migrator-59844c95c7-2bt9s\" (UID: \"e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.983125 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgkbb\" (UniqueName: \"kubernetes.io/projected/c54b345b-4399-418f-9bf2-a4c81e8c558b-kube-api-access-hgkbb\") pod \"ingress-canary-77m52\" (UID: \"c54b345b-4399-418f-9bf2-a4c81e8c558b\") " pod="openshift-ingress-canary/ingress-canary-77m52" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.988982 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.992094 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.992332 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.492297357 +0000 UTC m=+142.942012895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.992589 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/501c7fd2-e29f-48d8-9cf5-655af1e87402-metrics-tls\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.992711 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-certs\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.992773 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.992874 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-node-bootstrap-token\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.992917 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501c7fd2-e29f-48d8-9cf5-655af1e87402-config-volume\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.993062 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" event={"ID":"fcf134f5-5c52-447f-b5b7-d3fbe2aef728","Type":"ContainerStarted","Data":"0b8e3c7aecde741363cb6c4eefc968489315d0e6fadedfbbe6f5768893dd893e"} Oct 10 14:47:19 crc kubenswrapper[4788]: I1010 14:47:19.993824 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/501c7fd2-e29f-48d8-9cf5-655af1e87402-config-volume\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:19 crc kubenswrapper[4788]: E1010 14:47:19.994800 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.494781836 +0000 UTC m=+142.944497384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:19.998440 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8txtm"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.000167 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v225m\" (UniqueName: \"kubernetes.io/projected/6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078-kube-api-access-v225m\") pod \"machine-config-operator-74547568cd-hhq8v\" (UID: \"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.000195 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-certs\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.000630 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" event={"ID":"99fc4a35-5cf5-41a6-82a9-0814ac116d7d","Type":"ContainerStarted","Data":"42c9538ef4be62c4a365b2cc5d1f09ef75a06fb5b3e96375bc663696b347e168"} Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.003565 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/501c7fd2-e29f-48d8-9cf5-655af1e87402-metrics-tls\") pod \"dns-default-6xm66\" (UID: \"501c7fd2-e29f-48d8-9cf5-655af1e87402\") " pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.004541 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c3060c66-e983-43ae-b021-0a5c9b6dc99b-node-bootstrap-token\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.018307 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv7vk\" (UniqueName: \"kubernetes.io/projected/d8826b11-1579-4653-953e-0895bc1d16b5-kube-api-access-kv7vk\") pod \"console-f9d7485db-mgbhm\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.023468 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" Oct 10 14:47:20 crc kubenswrapper[4788]: W1010 14:47:20.035422 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb5d919_7db9_4efa_95ff_3bb5c5ea34a0.slice/crio-9607d46814f77ee55a8a5f9500a7b657f11889f52965de031d5dfad7012aa05e WatchSource:0}: Error finding container 9607d46814f77ee55a8a5f9500a7b657f11889f52965de031d5dfad7012aa05e: Status 404 returned error can't find the container with id 9607d46814f77ee55a8a5f9500a7b657f11889f52965de031d5dfad7012aa05e Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.035922 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9bw5d"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.041226 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-77m52" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.044279 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dngzx\" (UniqueName: \"kubernetes.io/projected/360b557a-fc75-4704-8cb2-fb9285d18566-kube-api-access-dngzx\") pod \"csi-hostpathplugin-xfl9r\" (UID: \"360b557a-fc75-4704-8cb2-fb9285d18566\") " pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.045681 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.050519 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:20 crc kubenswrapper[4788]: W1010 14:47:20.051818 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda46b6887_dafb_4cd1_a4f6_2a9602b84b89.slice/crio-faf23d6dfd0d8e459f5fce8415c8e258e6cc5ccfb85cd0af1a220d50a49fcae4 WatchSource:0}: Error finding container faf23d6dfd0d8e459f5fce8415c8e258e6cc5ccfb85cd0af1a220d50a49fcae4: Status 404 returned error can't find the container with id faf23d6dfd0d8e459f5fce8415c8e258e6cc5ccfb85cd0af1a220d50a49fcae4 Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.060672 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztkhv\" (UniqueName: \"kubernetes.io/projected/8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb-kube-api-access-ztkhv\") pod \"service-ca-9c57cc56f-7zmwn\" (UID: \"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb\") " pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.062526 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.080505 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf4nc\" (UniqueName: \"kubernetes.io/projected/1ce4a5b7-557b-4b26-924c-2e246030b180-kube-api-access-hf4nc\") pod \"router-default-5444994796-99fw8\" (UID: \"1ce4a5b7-557b-4b26-924c-2e246030b180\") " pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.089730 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.092505 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-h28k6"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.094116 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.094609 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.594594413 +0000 UTC m=+143.044309961 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.098636 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt79c\" (UniqueName: \"kubernetes.io/projected/5b667bac-e257-4741-b4e7-38857403b391-kube-api-access-wt79c\") pod \"service-ca-operator-777779d784-wslmf\" (UID: \"5b667bac-e257-4741-b4e7-38857403b391\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:20 crc kubenswrapper[4788]: W1010 14:47:20.110025 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d2c7d67_2055_4e2b_9598_4438c3b702c1.slice/crio-86d8a9ba537f8559bcbbbf50e440a4c5c01ad9eed35afa92347b5c3d253483df WatchSource:0}: Error finding container 86d8a9ba537f8559bcbbbf50e440a4c5c01ad9eed35afa92347b5c3d253483df: Status 404 returned error can't find the container with id 86d8a9ba537f8559bcbbbf50e440a4c5c01ad9eed35afa92347b5c3d253483df Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.121339 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.122105 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gnrr\" (UniqueName: \"kubernetes.io/projected/ffc68803-aa9e-4151-abc6-6d6a3eef7a8b-kube-api-access-8gnrr\") pod \"multus-admission-controller-857f4d67dd-8vx6h\" (UID: \"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.138312 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlss6\" (UniqueName: \"kubernetes.io/projected/2686886b-00d1-4e24-8a84-f634c766fcca-kube-api-access-wlss6\") pod \"marketplace-operator-79b997595-nrzg2\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.139330 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m448m"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.154499 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.160336 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.162553 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/19cfebf6-d761-4be2-a146-082c2a71deba-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.185166 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfvb7\" (UniqueName: \"kubernetes.io/projected/e6875754-4b83-4a13-b8e6-329300988b54-kube-api-access-nfvb7\") pod \"catalog-operator-68c6474976-cpvkv\" (UID: \"e6875754-4b83-4a13-b8e6-329300988b54\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.191405 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.195974 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.196364 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.696351884 +0000 UTC m=+143.146067432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.197300 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.198363 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lglr9\" (UniqueName: \"kubernetes.io/projected/7199dcec-2a4b-418b-ba9b-d04810aa6e70-kube-api-access-lglr9\") pod \"kube-storage-version-migrator-operator-b67b599dd-mrcc7\" (UID: \"7199dcec-2a4b-418b-ba9b-d04810aa6e70\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:20 crc kubenswrapper[4788]: W1010 14:47:20.201066 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10e304aa_8568_440e_99cc_f9081412645e.slice/crio-909f1e8746ae0b2a5cc01fcc05e5ac042ff869f55a13324be8167ad86aab8bbd WatchSource:0}: Error finding container 909f1e8746ae0b2a5cc01fcc05e5ac042ff869f55a13324be8167ad86aab8bbd: Status 404 returned error can't find the container with id 909f1e8746ae0b2a5cc01fcc05e5ac042ff869f55a13324be8167ad86aab8bbd Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.205396 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.207517 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.218005 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.230377 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgsks\" (UniqueName: \"kubernetes.io/projected/ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7-kube-api-access-mgsks\") pod \"package-server-manager-789f6589d5-2cft7\" (UID: \"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.232915 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.239345 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5lpl\" (UniqueName: \"kubernetes.io/projected/f0008884-e2e1-4229-ae6d-86bd85c7c301-kube-api-access-c5lpl\") pod \"olm-operator-6b444d44fb-lcsfw\" (UID: \"f0008884-e2e1-4229-ae6d-86bd85c7c301\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.250206 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.259243 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9x89\" (UniqueName: \"kubernetes.io/projected/aad02b5d-b3d9-4005-90d4-8e824d28a95d-kube-api-access-x9x89\") pod \"authentication-operator-69f744f599-459pb\" (UID: \"aad02b5d-b3d9-4005-90d4-8e824d28a95d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.263818 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.272091 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.283953 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjhpt\" (UniqueName: \"kubernetes.io/projected/e7ec1502-208c-48bb-b8ad-8cb5ab938da2-kube-api-access-gjhpt\") pod \"console-operator-58897d9998-cjsq9\" (UID: \"e7ec1502-208c-48bb-b8ad-8cb5ab938da2\") " pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.289561 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.296187 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.296768 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.296889 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.796857737 +0000 UTC m=+143.246573285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.297025 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.297669 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.797654116 +0000 UTC m=+143.247369654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.300083 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpwwz\" (UniqueName: \"kubernetes.io/projected/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-kube-api-access-jpwwz\") pod \"collect-profiles-29335125-24p8r\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.309750 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.325678 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc27468f-ca2c-4756-9d8f-d3b7a3325e3e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6ccmq\" (UID: \"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.333979 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.347443 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq9rs\" (UniqueName: \"kubernetes.io/projected/05e27a21-9c12-471b-b959-0a8a2046a371-kube-api-access-dq9rs\") pod \"machine-config-controller-84d6567774-8dhqw\" (UID: \"05e27a21-9c12-471b-b959-0a8a2046a371\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.365128 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khxqz\" (UniqueName: \"kubernetes.io/projected/f7897a5f-602b-4d6d-b266-4a80080d8671-kube-api-access-khxqz\") pod \"packageserver-d55dfcdfc-8tmcp\" (UID: \"f7897a5f-602b-4d6d-b266-4a80080d8671\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.372273 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.379928 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6gtc\" (UniqueName: \"kubernetes.io/projected/bb4a9174-7055-4a98-a3dc-790713348443-kube-api-access-v6gtc\") pod \"etcd-operator-b45778765-qvvxg\" (UID: \"bb4a9174-7055-4a98-a3dc-790713348443\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.398773 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.399447 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:20.899419758 +0000 UTC m=+143.349135426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.401414 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f007d78-6f96-4523-8ae6-7d03955f0743-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-22mm5\" (UID: \"2f007d78-6f96-4523-8ae6-7d03955f0743\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.419983 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f28vg\" (UniqueName: \"kubernetes.io/projected/19cfebf6-d761-4be2-a146-082c2a71deba-kube-api-access-f28vg\") pod \"ingress-operator-5b745b69d9-6w9gx\" (UID: \"19cfebf6-d761-4be2-a146-082c2a71deba\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.429695 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.450812 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wrlf\" (UniqueName: \"kubernetes.io/projected/6b2c9335-c98e-4808-b9cf-ecac89cd1a99-kube-api-access-9wrlf\") pod \"control-plane-machine-set-operator-78cbb6b69f-cgbfx\" (UID: \"6b2c9335-c98e-4808-b9cf-ecac89cd1a99\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.460406 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4n2l\" (UniqueName: \"kubernetes.io/projected/c3060c66-e983-43ae-b021-0a5c9b6dc99b-kube-api-access-r4n2l\") pod \"machine-config-server-2g9kl\" (UID: \"c3060c66-e983-43ae-b021-0a5c9b6dc99b\") " pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.468950 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.484500 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.502426 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9pz2\" (UniqueName: \"kubernetes.io/projected/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-kube-api-access-w9pz2\") pod \"oauth-openshift-558db77b4-5j9sp\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.505292 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.505506 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.511009 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.511818 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.011794734 +0000 UTC m=+143.461510472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.512944 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.520118 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.526410 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.546448 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.556322 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.585944 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.603048 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.604223 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6xm66"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.616858 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-77m52"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.624770 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.625674 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.125650978 +0000 UTC m=+143.575366526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.661645 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2g9kl" Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.672185 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.711922 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:20 crc kubenswrapper[4788]: W1010 14:47:20.722898 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01c39c4c_1b9c_464e_aba0_434acf814e34.slice/crio-9f23a2a72bb1bc8ee98956fe6efb93ff279795013dc0680d8f684c2a376324af WatchSource:0}: Error finding container 9f23a2a72bb1bc8ee98956fe6efb93ff279795013dc0680d8f684c2a376324af: Status 404 returned error can't find the container with id 9f23a2a72bb1bc8ee98956fe6efb93ff279795013dc0680d8f684c2a376324af Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.727235 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.727542 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.227527793 +0000 UTC m=+143.677243341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: W1010 14:47:20.764198 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc54b345b_4399_418f_9bf2_a4c81e8c558b.slice/crio-cbd80beae0a8efb64d3cba19ad2c0c4568a8a381a6cf14d41aa5080baa47055b WatchSource:0}: Error finding container cbd80beae0a8efb64d3cba19ad2c0c4568a8a381a6cf14d41aa5080baa47055b: Status 404 returned error can't find the container with id cbd80beae0a8efb64d3cba19ad2c0c4568a8a381a6cf14d41aa5080baa47055b Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.827756 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.828364 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.828686 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.32866531 +0000 UTC m=+143.778380858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.828760 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.829092 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.3290852 +0000 UTC m=+143.778800748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.835352 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7xj4z"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.838585 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mgbhm"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.839878 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.863553 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nrzg2"] Oct 10 14:47:20 crc kubenswrapper[4788]: I1010 14:47:20.931057 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:20 crc kubenswrapper[4788]: E1010 14:47:20.931510 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.431495738 +0000 UTC m=+143.881211286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:20 crc kubenswrapper[4788]: W1010 14:47:20.938803 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3060c66_e983_43ae_b021_0a5c9b6dc99b.slice/crio-863494bcea95c2d3153b5ec828b2984055aae889edb942c14a888b533681bbda WatchSource:0}: Error finding container 863494bcea95c2d3153b5ec828b2984055aae889edb942c14a888b533681bbda: Status 404 returned error can't find the container with id 863494bcea95c2d3153b5ec828b2984055aae889edb942c14a888b533681bbda Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.025252 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-99fw8" event={"ID":"1ce4a5b7-557b-4b26-924c-2e246030b180","Type":"ContainerStarted","Data":"1d9bd55afde870b60496e6e3f17543cdfc866d448bb6e9f8246483e893c6ab27"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.033120 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.033618 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.533597678 +0000 UTC m=+143.983313226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.033701 4788 generic.go:334] "Generic (PLEG): container finished" podID="cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0" containerID="51c3cdfa11a43b1138373e94e0c66d1f53be68bea7b81a1f83e895c9b1104440" exitCode=0 Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.034244 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" event={"ID":"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0","Type":"ContainerDied","Data":"51c3cdfa11a43b1138373e94e0c66d1f53be68bea7b81a1f83e895c9b1104440"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.034299 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" event={"ID":"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0","Type":"ContainerStarted","Data":"9607d46814f77ee55a8a5f9500a7b657f11889f52965de031d5dfad7012aa05e"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.036792 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" event={"ID":"51751bfc-73b2-4968-bbf7-a213e109b40a","Type":"ContainerStarted","Data":"f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.036855 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" event={"ID":"51751bfc-73b2-4968-bbf7-a213e109b40a","Type":"ContainerStarted","Data":"ab33f62f4ff25bd46501cc6f50e1f4b8c8e6a72617bbb6957f253160d1ee138c"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.037813 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.039296 4788 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-rnqfq container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.039459 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" podUID="51751bfc-73b2-4968-bbf7-a213e109b40a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.042271 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" event={"ID":"10e304aa-8568-440e-99cc-f9081412645e","Type":"ContainerStarted","Data":"bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.042303 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" event={"ID":"10e304aa-8568-440e-99cc-f9081412645e","Type":"ContainerStarted","Data":"909f1e8746ae0b2a5cc01fcc05e5ac042ff869f55a13324be8167ad86aab8bbd"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.043057 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.045549 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" event={"ID":"c389c203-6864-4423-bb2a-da5693e052dc","Type":"ContainerStarted","Data":"6bc0f1655e6ed840eff98f82be4625340516c4b55cc316668055cb8dcd0ac8ba"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.046560 4788 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-m448m container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.046620 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" podUID="10e304aa-8568-440e-99cc-f9081412645e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.048104 4788 generic.go:334] "Generic (PLEG): container finished" podID="fcf134f5-5c52-447f-b5b7-d3fbe2aef728" containerID="af553fedfff246e49e255312bdddd83973b5cea6f4d63a4f90c55fe8476337a4" exitCode=0 Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.048864 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" event={"ID":"fcf134f5-5c52-447f-b5b7-d3fbe2aef728","Type":"ContainerDied","Data":"af553fedfff246e49e255312bdddd83973b5cea6f4d63a4f90c55fe8476337a4"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.054459 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2g9kl" event={"ID":"c3060c66-e983-43ae-b021-0a5c9b6dc99b","Type":"ContainerStarted","Data":"863494bcea95c2d3153b5ec828b2984055aae889edb942c14a888b533681bbda"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.067516 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" event={"ID":"a46b6887-dafb-4cd1-a4f6-2a9602b84b89","Type":"ContainerStarted","Data":"f073f8ed13fd168471cb5ca53140d19ec680b7d62c69475fcfe472d8c24df8a0"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.067596 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" event={"ID":"a46b6887-dafb-4cd1-a4f6-2a9602b84b89","Type":"ContainerStarted","Data":"faf23d6dfd0d8e459f5fce8415c8e258e6cc5ccfb85cd0af1a220d50a49fcae4"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.070633 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" event={"ID":"2821f67e-912c-48d7-8c95-47b29949afe9","Type":"ContainerStarted","Data":"30bbc08b6b86210541a2feb3ced10af3c6a8aa5721e3db18dc2fc77ababa7ec1"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.073307 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xm66" event={"ID":"501c7fd2-e29f-48d8-9cf5-655af1e87402","Type":"ContainerStarted","Data":"74adb142457fd335b39b43d7b5665d0cb2099734b3bc20a1a9148760a87686d3"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.077500 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-77m52" event={"ID":"c54b345b-4399-418f-9bf2-a4c81e8c558b","Type":"ContainerStarted","Data":"cbd80beae0a8efb64d3cba19ad2c0c4568a8a381a6cf14d41aa5080baa47055b"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.081072 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-h28k6" event={"ID":"7d2c7d67-2055-4e2b-9598-4438c3b702c1","Type":"ContainerStarted","Data":"fc2238f4bbd77dfc6f8374e1743c9f94f37988aaedef24dfa28a007398073b51"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.081153 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-h28k6" event={"ID":"7d2c7d67-2055-4e2b-9598-4438c3b702c1","Type":"ContainerStarted","Data":"86d8a9ba537f8559bcbbbf50e440a4c5c01ad9eed35afa92347b5c3d253483df"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.081549 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-h28k6" Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.083102 4788 patch_prober.go:28] interesting pod/downloads-7954f5f757-h28k6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.083196 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h28k6" podUID="7d2c7d67-2055-4e2b-9598-4438c3b702c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.086993 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" event={"ID":"99fc4a35-5cf5-41a6-82a9-0814ac116d7d","Type":"ContainerStarted","Data":"afa6ba2cc4345ed0d0df68fbf10915e195b2348bd66393ff2921bf3be1b9a148"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.087272 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" event={"ID":"99fc4a35-5cf5-41a6-82a9-0814ac116d7d","Type":"ContainerStarted","Data":"2fdaca1407baec11c2f262a43f0fe645817c9eda81cd5e46d4376b7023ffb9ce"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.097413 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" event={"ID":"01c39c4c-1b9c-464e-aba0-434acf814e34","Type":"ContainerStarted","Data":"9f23a2a72bb1bc8ee98956fe6efb93ff279795013dc0680d8f684c2a376324af"} Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.136231 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.136658 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.636618581 +0000 UTC m=+144.086334289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.244599 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.250647 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.750614837 +0000 UTC m=+144.200330385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.353677 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.354257 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.854235243 +0000 UTC m=+144.303950801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.455932 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.457129 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:21.957103483 +0000 UTC m=+144.406819031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.531750 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8vx6h"] Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.559081 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.559682 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.059658414 +0000 UTC m=+144.509373962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.662722 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.663968 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.163947896 +0000 UTC m=+144.613663444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.769100 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.769611 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.269590582 +0000 UTC m=+144.719306130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.870320 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.871046 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.371034787 +0000 UTC m=+144.820750335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.932433 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wslmf"] Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.950500 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s"] Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.972953 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:21 crc kubenswrapper[4788]: E1010 14:47:21.973542 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.473517886 +0000 UTC m=+144.923233434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:21 crc kubenswrapper[4788]: I1010 14:47:21.978687 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.005764 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7zmwn"] Oct 10 14:47:22 crc kubenswrapper[4788]: W1010 14:47:22.007504 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b667bac_e257_4741_b4e7_38857403b391.slice/crio-982b9d248fac5629bd39333cebed968fcc7736cd07b7972f159ef8c50918dcc2 WatchSource:0}: Error finding container 982b9d248fac5629bd39333cebed968fcc7736cd07b7972f159ef8c50918dcc2: Status 404 returned error can't find the container with id 982b9d248fac5629bd39333cebed968fcc7736cd07b7972f159ef8c50918dcc2 Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.045514 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.045576 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xfl9r"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.063476 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.075558 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.075967 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.575954775 +0000 UTC m=+145.025670313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.169842 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.170663 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cjsq9"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.176462 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.176938 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.676918719 +0000 UTC m=+145.126634267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.198371 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" event={"ID":"fcf134f5-5c52-447f-b5b7-d3fbe2aef728","Type":"ContainerStarted","Data":"e5c31ed1bb960161a1f7ab68c70cb0ca29b6efb25a8a8fb2786900cb39c35709"} Oct 10 14:47:22 crc kubenswrapper[4788]: W1010 14:47:22.202805 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6daf8c88_e9e8_48cf_acb8_c7a9f8cd9078.slice/crio-749b7aad969ca27ea002fce97a2a1472159c47454ce68531dfedbdf2bade1085 WatchSource:0}: Error finding container 749b7aad969ca27ea002fce97a2a1472159c47454ce68531dfedbdf2bade1085: Status 404 returned error can't find the container with id 749b7aad969ca27ea002fce97a2a1472159c47454ce68531dfedbdf2bade1085 Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.209862 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" event={"ID":"cfb5d919-7db9-4efa-95ff-3bb5c5ea34a0","Type":"ContainerStarted","Data":"54cab48219954d2cd53ab571cbcd60dab5e9835e64536171fa49675b3d49f080"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.210040 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.213916 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw"] Oct 10 14:47:22 crc kubenswrapper[4788]: W1010 14:47:22.214873 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7ec1502_208c_48bb_b8ad_8cb5ab938da2.slice/crio-a5a5f122dfa312541c01ca56ba36a3545339d168b388523429c3b408c9046ca8 WatchSource:0}: Error finding container a5a5f122dfa312541c01ca56ba36a3545339d168b388523429c3b408c9046ca8: Status 404 returned error can't find the container with id a5a5f122dfa312541c01ca56ba36a3545339d168b388523429c3b408c9046ca8 Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.216961 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" event={"ID":"a46b6887-dafb-4cd1-a4f6-2a9602b84b89","Type":"ContainerStarted","Data":"1ae74a5c26fd4cd8f4b14b230969cd97b786b6215dec1285f415535a16e2d18c"} Oct 10 14:47:22 crc kubenswrapper[4788]: W1010 14:47:22.218357 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19cfebf6_d761_4be2_a146_082c2a71deba.slice/crio-78c09d86552528bd3959cc050a351f5043fbe7a7cef6f86a16a7538e8cd716be WatchSource:0}: Error finding container 78c09d86552528bd3959cc050a351f5043fbe7a7cef6f86a16a7538e8cd716be: Status 404 returned error can't find the container with id 78c09d86552528bd3959cc050a351f5043fbe7a7cef6f86a16a7538e8cd716be Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.219243 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" event={"ID":"ee1391a9-17fa-48f0-9145-6356860930f8","Type":"ContainerStarted","Data":"e85c0490fd568aaa8ed75cd47a11a44fa7ad69a744f31831efb870936556334b"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.220251 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.225930 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" event={"ID":"e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7","Type":"ContainerStarted","Data":"1bc10ceae5209ac09943c7ae6f8ef7ddf16b3174ea8053c610edda2acf3586fc"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.230344 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" event={"ID":"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb","Type":"ContainerStarted","Data":"2989b55f2acb9890f121ebd79710a6f6b4638cf1d88dc4d32fe9502f800496f2"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.278133 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.279396 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.779369707 +0000 UTC m=+145.229085245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.308760 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" event={"ID":"cedd44c5-a11e-48e9-b61f-04c087777fc9","Type":"ContainerStarted","Data":"e1d35cba3364d81dd246a58e632328ae4d445c638dd40d6f966bb7a9d51b99e6"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.308794 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.308811 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.308821 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgbhm" event={"ID":"d8826b11-1579-4653-953e-0895bc1d16b5","Type":"ContainerStarted","Data":"338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.308832 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgbhm" event={"ID":"d8826b11-1579-4653-953e-0895bc1d16b5","Type":"ContainerStarted","Data":"92e63c9af9921dd9cb2e51ee57fa4720cf93e2197e29970899340b7861883b01"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.308842 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-459pb"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.308860 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" event={"ID":"47d58cb0-69cb-485b-b1e1-c857093a4c3c","Type":"ContainerStarted","Data":"1fe7f3bff533a4c6ee26cbab9fe04d65125a65dc9d421fb43a80fb1719081be5"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.308869 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" event={"ID":"47d58cb0-69cb-485b-b1e1-c857093a4c3c","Type":"ContainerStarted","Data":"79c052831511f35968a3df204849fef3241057c7480f37f3eff940b1268cce20"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.328570 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" event={"ID":"01c39c4c-1b9c-464e-aba0-434acf814e34","Type":"ContainerStarted","Data":"f34eab4df23be049aafbf1a2192e7b8570862205f6fecb0f6b8c0b78276a5f95"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.338049 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-h28k6" podStartSLOduration=125.338029645 podStartE2EDuration="2m5.338029645s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.297914062 +0000 UTC m=+144.747629610" watchObservedRunningTime="2025-10-10 14:47:22.338029645 +0000 UTC m=+144.787745193" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.343970 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" event={"ID":"360b557a-fc75-4704-8cb2-fb9285d18566","Type":"ContainerStarted","Data":"11deae7f03e018602f0db6c5e45cfd79d93f5febf0c664e6eedeabef10299cc6"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.348420 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" event={"ID":"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4","Type":"ContainerStarted","Data":"822a7fddae931d52b2c18a34e0b13a491b032132cadbd40ff2cb2892f042d7e6"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.348462 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" event={"ID":"2e6e895e-b3ff-4a54-9147-0e5e962bbdc4","Type":"ContainerStarted","Data":"1a127797810863c89e16b62e05f6acf5f2ea12a0d1eb556558eff5320a8e9db5"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.358030 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xm66" event={"ID":"501c7fd2-e29f-48d8-9cf5-655af1e87402","Type":"ContainerStarted","Data":"246d7eafe1c98ab8eaa2b63ced98171245c004cc3ecbcd7831206bfb003a2387"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.363369 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" event={"ID":"5b667bac-e257-4741-b4e7-38857403b391","Type":"ContainerStarted","Data":"982b9d248fac5629bd39333cebed968fcc7736cd07b7972f159ef8c50918dcc2"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.374290 4788 generic.go:334] "Generic (PLEG): container finished" podID="c389c203-6864-4423-bb2a-da5693e052dc" containerID="46b8ce7cdec35e899ae613998c8c2609bd22992657b3348b128d893aa7372a5b" exitCode=0 Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.374424 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" event={"ID":"c389c203-6864-4423-bb2a-da5693e052dc","Type":"ContainerDied","Data":"46b8ce7cdec35e899ae613998c8c2609bd22992657b3348b128d893aa7372a5b"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.384890 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.386450 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:22.886425866 +0000 UTC m=+145.336141414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.388015 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-99fw8" event={"ID":"1ce4a5b7-557b-4b26-924c-2e246030b180","Type":"ContainerStarted","Data":"76b5d8ef17b4e973e7abf1644e595afe67e81a7740db895c7603779f06755a51"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.400418 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" podStartSLOduration=125.400391901 podStartE2EDuration="2m5.400391901s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.337593394 +0000 UTC m=+144.787308942" watchObservedRunningTime="2025-10-10 14:47:22.400391901 +0000 UTC m=+144.850107449" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.406846 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2g9kl" event={"ID":"c3060c66-e983-43ae-b021-0a5c9b6dc99b","Type":"ContainerStarted","Data":"51879f6040d3911a68a432328645f948693ddb5b49a1e8a6e4fdcb61f381e85e"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.432079 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mgbhm" podStartSLOduration=125.432061512 podStartE2EDuration="2m5.432061512s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.404233144 +0000 UTC m=+144.853948702" watchObservedRunningTime="2025-10-10 14:47:22.432061512 +0000 UTC m=+144.881777050" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.433238 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" podStartSLOduration=124.43323135 podStartE2EDuration="2m4.43323135s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.431468268 +0000 UTC m=+144.881183816" watchObservedRunningTime="2025-10-10 14:47:22.43323135 +0000 UTC m=+144.882946898" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.437025 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.443968 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" event={"ID":"2821f67e-912c-48d7-8c95-47b29949afe9","Type":"ContainerStarted","Data":"b92bf4e418c8703a5468fd08b8cf135d7ef0141c0d228e740f76d09fa9f670d0"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.474771 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5j9sp"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.476346 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.529526 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-9bw5d" podStartSLOduration=124.52948793 podStartE2EDuration="2m4.52948793s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.476134349 +0000 UTC m=+144.925849897" watchObservedRunningTime="2025-10-10 14:47:22.52948793 +0000 UTC m=+144.979203478" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.532268 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.536386 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.036371005 +0000 UTC m=+145.486086553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.541399 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.546664 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.552804 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qvvxg"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.564889 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" podStartSLOduration=124.564873429 podStartE2EDuration="2m4.564873429s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.564666225 +0000 UTC m=+145.014381773" watchObservedRunningTime="2025-10-10 14:47:22.564873429 +0000 UTC m=+145.014588977" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.565427 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" event={"ID":"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b","Type":"ContainerStarted","Data":"3393d9ecdda785deebbb81d1dd0ad95683adcb2dc97fbc572d3ba1f5415dfa0b"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.573520 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5"] Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.591985 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwlxm" podStartSLOduration=125.591966139 podStartE2EDuration="2m5.591966139s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.589020079 +0000 UTC m=+145.038735647" watchObservedRunningTime="2025-10-10 14:47:22.591966139 +0000 UTC m=+145.041681687" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.593668 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-77m52" event={"ID":"c54b345b-4399-418f-9bf2-a4c81e8c558b","Type":"ContainerStarted","Data":"924a19f58a98da324cb98e2ac92708c653c9777847875d977f63ef7db1185507"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.602727 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" event={"ID":"2686886b-00d1-4e24-8a84-f634c766fcca","Type":"ContainerStarted","Data":"bd354c6be945783100ba01d4b72112ac9d919792a526513c514770caf3210e61"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.603281 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" event={"ID":"2686886b-00d1-4e24-8a84-f634c766fcca","Type":"ContainerStarted","Data":"955e6795adf937975d3cb7afa4aa8c1bad75561298ea0540caa87244da9ee037"} Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.604343 4788 patch_prober.go:28] interesting pod/downloads-7954f5f757-h28k6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.604386 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h28k6" podUID="7d2c7d67-2055-4e2b-9598-4438c3b702c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.606599 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.608181 4788 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nrzg2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/healthz\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.608210 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" podUID="2686886b-00d1-4e24-8a84-f634c766fcca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.16:8080/healthz\": dial tcp 10.217.0.16:8080: connect: connection refused" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.620988 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.621535 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.624102 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-99fw8" podStartSLOduration=124.62408557 podStartE2EDuration="2m4.62408557s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.622776569 +0000 UTC m=+145.072492117" watchObservedRunningTime="2025-10-10 14:47:22.62408557 +0000 UTC m=+145.073801118" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.634759 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.636117 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.136074918 +0000 UTC m=+145.585790466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.637287 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.640476 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.140462544 +0000 UTC m=+145.590178092 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.669768 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h7d2s" podStartSLOduration=124.669751017 podStartE2EDuration="2m4.669751017s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.666788816 +0000 UTC m=+145.116504364" watchObservedRunningTime="2025-10-10 14:47:22.669751017 +0000 UTC m=+145.119466565" Oct 10 14:47:22 crc kubenswrapper[4788]: W1010 14:47:22.702465 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc27468f_ca2c_4756_9d8f_d3b7a3325e3e.slice/crio-8e69ede30e685110fde6735c0a54d00ae5647a332543339f2661addc8b3b23b2 WatchSource:0}: Error finding container 8e69ede30e685110fde6735c0a54d00ae5647a332543339f2661addc8b3b23b2: Status 404 returned error can't find the container with id 8e69ede30e685110fde6735c0a54d00ae5647a332543339f2661addc8b3b23b2 Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.738879 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.739006 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.238988028 +0000 UTC m=+145.688703576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.739642 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.740012 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.240003623 +0000 UTC m=+145.689719171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.744481 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2g9kl" podStartSLOduration=5.744466629 podStartE2EDuration="5.744466629s" podCreationTimestamp="2025-10-10 14:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.704436208 +0000 UTC m=+145.154151776" watchObservedRunningTime="2025-10-10 14:47:22.744466629 +0000 UTC m=+145.194182167" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.794663 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hvvxp" podStartSLOduration=124.794643484 podStartE2EDuration="2m4.794643484s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.788358863 +0000 UTC m=+145.238074411" watchObservedRunningTime="2025-10-10 14:47:22.794643484 +0000 UTC m=+145.244359032" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.829904 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8sb2v" podStartSLOduration=124.82988762 podStartE2EDuration="2m4.82988762s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.828945947 +0000 UTC m=+145.278661495" watchObservedRunningTime="2025-10-10 14:47:22.82988762 +0000 UTC m=+145.279603168" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.840397 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.840787 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.340772321 +0000 UTC m=+145.790487869 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.869807 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" podStartSLOduration=124.869789058 podStartE2EDuration="2m4.869789058s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.869447289 +0000 UTC m=+145.319162837" watchObservedRunningTime="2025-10-10 14:47:22.869789058 +0000 UTC m=+145.319504606" Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.946222 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:22 crc kubenswrapper[4788]: E1010 14:47:22.946601 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.44658803 +0000 UTC m=+145.896303578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:22 crc kubenswrapper[4788]: I1010 14:47:22.988184 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-77m52" podStartSLOduration=5.988126957 podStartE2EDuration="5.988126957s" podCreationTimestamp="2025-10-10 14:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:22.945812152 +0000 UTC m=+145.395527700" watchObservedRunningTime="2025-10-10 14:47:22.988126957 +0000 UTC m=+145.437842505" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.047909 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.048848 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.548806513 +0000 UTC m=+145.998522061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.156800 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.157256 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.657238066 +0000 UTC m=+146.106953614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.258270 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.258905 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.758888435 +0000 UTC m=+146.208603983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.359507 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.359847 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.859831059 +0000 UTC m=+146.309546597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.375545 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.388482 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:23 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:23 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:23 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.388547 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.464208 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.464844 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:23.964790108 +0000 UTC m=+146.414505656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.570723 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.571325 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.071310084 +0000 UTC m=+146.521025632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.641792 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" event={"ID":"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7","Type":"ContainerStarted","Data":"683ec2dca0c2520c9a33c5066b19428348f9672d57f0f31134b72a698bdaee17"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.648327 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" event={"ID":"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b","Type":"ContainerStarted","Data":"0d1ae0e5627c8c1e1e804821ab81bc36cad17e891e29d4496f51faa01ca15a6e"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.677878 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.678330 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.178315042 +0000 UTC m=+146.628030580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.713414 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" event={"ID":"8cdf6c37-aaa9-4d9a-acfa-3158f1258ddb","Type":"ContainerStarted","Data":"54ba5cf710854ba117292f641aba809fc692c6cb36db98d3b75e0be45a86b1e7"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.735436 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" event={"ID":"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078","Type":"ContainerStarted","Data":"c297fa462d67ba6b3bee1a6fd886bec00393962d70735a6bb3131bf4417cdfa9"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.735484 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" event={"ID":"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078","Type":"ContainerStarted","Data":"749b7aad969ca27ea002fce97a2a1472159c47454ce68531dfedbdf2bade1085"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.736852 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" event={"ID":"aad02b5d-b3d9-4005-90d4-8e824d28a95d","Type":"ContainerStarted","Data":"c07aa2f6f489eeab43ccbf9aba372a25bef953edc6550516aac73d932b6c2b1b"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.738081 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" event={"ID":"cedd44c5-a11e-48e9-b61f-04c087777fc9","Type":"ContainerStarted","Data":"6be111816b55596e9f91c304d097b559d70ea058b1d53a1b3fe415a0048256ab"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.761460 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-7zmwn" podStartSLOduration=125.761447067 podStartE2EDuration="2m5.761447067s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:23.759577092 +0000 UTC m=+146.209292650" watchObservedRunningTime="2025-10-10 14:47:23.761447067 +0000 UTC m=+146.211162615" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.779004 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" event={"ID":"e7ec1502-208c-48bb-b8ad-8cb5ab938da2","Type":"ContainerStarted","Data":"45bbdf6b028cacc0c5e54280465944c6720786e32179449cc6c95a4a1d5a0f91"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.779045 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" event={"ID":"e7ec1502-208c-48bb-b8ad-8cb5ab938da2","Type":"ContainerStarted","Data":"a5a5f122dfa312541c01ca56ba36a3545339d168b388523429c3b408c9046ca8"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.779874 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.780281 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.780591 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.280579596 +0000 UTC m=+146.730295144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.787561 4788 patch_prober.go:28] interesting pod/console-operator-58897d9998-cjsq9 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.787626 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" podUID="e7ec1502-208c-48bb-b8ad-8cb5ab938da2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.822704 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" podStartSLOduration=126.822686417 podStartE2EDuration="2m6.822686417s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:23.82195464 +0000 UTC m=+146.271670188" watchObservedRunningTime="2025-10-10 14:47:23.822686417 +0000 UTC m=+146.272401965" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.842216 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" event={"ID":"e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7","Type":"ContainerStarted","Data":"71e38cb8ad688b0c0b7e4b70ad0ed2d7e7b17cc515aa39fef6f46f1ec7c705ba"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.872013 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" podStartSLOduration=126.87200039 podStartE2EDuration="2m6.87200039s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:23.869435689 +0000 UTC m=+146.319151237" watchObservedRunningTime="2025-10-10 14:47:23.87200039 +0000 UTC m=+146.321715938" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.882358 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.883390 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.383360094 +0000 UTC m=+146.833075722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.922598 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" event={"ID":"2f007d78-6f96-4523-8ae6-7d03955f0743","Type":"ContainerStarted","Data":"780d663e4850baef4642e54f01015113bccb18dde1b8306252686f89a8d3c287"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.930041 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" event={"ID":"fcf134f5-5c52-447f-b5b7-d3fbe2aef728","Type":"ContainerStarted","Data":"7c7c980b545a7f1da43a3740bae4bfff297e16016c8169d5c1dea727e9163910"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.946608 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" event={"ID":"f7897a5f-602b-4d6d-b266-4a80080d8671","Type":"ContainerStarted","Data":"dc6af03086deb53d3c74c92b530c935eb221e585096b7adf29dd9c3ab8eb3c55"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.958429 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" podStartSLOduration=125.958412494 podStartE2EDuration="2m5.958412494s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:23.957948163 +0000 UTC m=+146.407663711" watchObservedRunningTime="2025-10-10 14:47:23.958412494 +0000 UTC m=+146.408128042" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.960432 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" event={"ID":"6b2c9335-c98e-4808-b9cf-ecac89cd1a99","Type":"ContainerStarted","Data":"c7b4a846d6f3b59c3dbca2b990544532bb03504ab932c3bf1bc40eb9783d4ce8"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.962333 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" event={"ID":"bb4a9174-7055-4a98-a3dc-790713348443","Type":"ContainerStarted","Data":"9f5f2243304a8fb9ba08ca58950e336fd6e2d91655c9097fdbcb5467d61a0abe"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.967566 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xm66" event={"ID":"501c7fd2-e29f-48d8-9cf5-655af1e87402","Type":"ContainerStarted","Data":"f38400fc7ff478dbfa44a18168afbb427470ac561d6e8130ce557701e0d4721a"} Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.968250 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.985026 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:23 crc kubenswrapper[4788]: E1010 14:47:23.986958 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.486946669 +0000 UTC m=+146.936662217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:23 crc kubenswrapper[4788]: I1010 14:47:23.989291 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" event={"ID":"ee1391a9-17fa-48f0-9145-6356860930f8","Type":"ContainerStarted","Data":"41640c963c736ef6908eaa70b50d81d7248e26c7341051787c848e84fbb44b9e"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.008337 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" event={"ID":"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e","Type":"ContainerStarted","Data":"8e69ede30e685110fde6735c0a54d00ae5647a332543339f2661addc8b3b23b2"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.030585 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" event={"ID":"19cfebf6-d761-4be2-a146-082c2a71deba","Type":"ContainerStarted","Data":"78c09d86552528bd3959cc050a351f5043fbe7a7cef6f86a16a7538e8cd716be"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.042260 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" podStartSLOduration=126.042243306 podStartE2EDuration="2m6.042243306s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.040609987 +0000 UTC m=+146.490325535" watchObservedRunningTime="2025-10-10 14:47:24.042243306 +0000 UTC m=+146.491958854" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.055062 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" event={"ID":"e6875754-4b83-4a13-b8e6-329300988b54","Type":"ContainerStarted","Data":"1838c99165fa79cf79bd0f0d222ba72de4d008ecefd42774c8eac9f442d17554"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.055100 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" event={"ID":"e6875754-4b83-4a13-b8e6-329300988b54","Type":"ContainerStarted","Data":"b9eb0be2228131f21ac03222e581822cdb7418a76b3fa9f626822e2cc47861dd"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.055443 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.076358 4788 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-cpvkv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.076411 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" podUID="e6875754-4b83-4a13-b8e6-329300988b54" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.093316 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.093452 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.593428545 +0000 UTC m=+147.043144093 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.093527 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.094698 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.594686645 +0000 UTC m=+147.044402193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.126449 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" event={"ID":"05e27a21-9c12-471b-b959-0a8a2046a371","Type":"ContainerStarted","Data":"b4ba9560f6c630aa7efe01327e1321355730c8264b77a4e2f3da141251b4f0e7"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.149226 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" event={"ID":"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05","Type":"ContainerStarted","Data":"f3fed2cc75e0676d281b6a80c6c408004d67492a4d370ab1492de4079c03ee26"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.168591 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" event={"ID":"f0008884-e2e1-4229-ae6d-86bd85c7c301","Type":"ContainerStarted","Data":"b6aa1f1e278ea4ac52835db240e49c0747ee5f61b382f8cabe2104cdc4fedf1e"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.168633 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" event={"ID":"f0008884-e2e1-4229-ae6d-86bd85c7c301","Type":"ContainerStarted","Data":"6696c39808062a87b9c82abb834be5025932c0c9f0b3b5f3bba7f23ae82e4769"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.169472 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.181295 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6xm66" podStartSLOduration=7.181277963 podStartE2EDuration="7.181277963s" podCreationTimestamp="2025-10-10 14:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.140393722 +0000 UTC m=+146.590109340" watchObservedRunningTime="2025-10-10 14:47:24.181277963 +0000 UTC m=+146.630993511" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.191309 4788 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-lcsfw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.191368 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" podUID="f0008884-e2e1-4229-ae6d-86bd85c7c301" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.193759 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" event={"ID":"47d58cb0-69cb-485b-b1e1-c857093a4c3c","Type":"ContainerStarted","Data":"b8a6822ea8d7e9417d39a7e4e05386fc6ddf406aa2aa2eb195323e414af4bbb8"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.195884 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.197064 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.697048622 +0000 UTC m=+147.146764170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.241104 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k74dd" podStartSLOduration=126.241065598 podStartE2EDuration="2m6.241065598s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.182452101 +0000 UTC m=+146.632167649" watchObservedRunningTime="2025-10-10 14:47:24.241065598 +0000 UTC m=+146.690781146" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.244301 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" podStartSLOduration=126.244211293 podStartE2EDuration="2m6.244211293s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.239802688 +0000 UTC m=+146.689518236" watchObservedRunningTime="2025-10-10 14:47:24.244211293 +0000 UTC m=+146.693926851" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.294305 4788 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nrzg2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/healthz\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.294363 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" podUID="2686886b-00d1-4e24-8a84-f634c766fcca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.16:8080/healthz\": dial tcp 10.217.0.16:8080: connect: connection refused" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.298276 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.337468 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.837431561 +0000 UTC m=+147.287147119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.350772 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" podStartSLOduration=126.35074312 podStartE2EDuration="2m6.35074312s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.343583968 +0000 UTC m=+146.793299516" watchObservedRunningTime="2025-10-10 14:47:24.35074312 +0000 UTC m=+146.800458668" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.366264 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" podStartSLOduration=126.366242542 podStartE2EDuration="2m6.366242542s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.27655703 +0000 UTC m=+146.726272578" watchObservedRunningTime="2025-10-10 14:47:24.366242542 +0000 UTC m=+146.815958090" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.387696 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7xj4z" podStartSLOduration=126.387667257 podStartE2EDuration="2m6.387667257s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.387204095 +0000 UTC m=+146.836919643" watchObservedRunningTime="2025-10-10 14:47:24.387667257 +0000 UTC m=+146.837382805" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.394680 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:24 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:24 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:24 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.395088 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.405259 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.407548 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:24.907516333 +0000 UTC m=+147.357231891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.444899 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" podStartSLOduration=126.44488257 podStartE2EDuration="2m6.44488257s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.416381026 +0000 UTC m=+146.866096574" watchObservedRunningTime="2025-10-10 14:47:24.44488257 +0000 UTC m=+146.894598118" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.446134 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" event={"ID":"7199dcec-2a4b-418b-ba9b-d04810aa6e70","Type":"ContainerStarted","Data":"95410be4f668066fb504feefd878f6e3c4dc5c665d4bc3f302b9f6440134f5cb"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.446214 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mrcc7" event={"ID":"7199dcec-2a4b-418b-ba9b-d04810aa6e70","Type":"ContainerStarted","Data":"349596e13cec71f098faed526264735bd63dc39924636d7a4310610d0048b0b3"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.446231 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wslmf" event={"ID":"5b667bac-e257-4741-b4e7-38857403b391","Type":"ContainerStarted","Data":"54aee273805b4cbf60da88a1efba16284fd2f81f2381885232907c28f9021d66"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.446245 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" event={"ID":"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a","Type":"ContainerStarted","Data":"1f325e4b115eab39402ec0157557788fe7f0bcddab951f36d250ed67a0fb4dc3"} Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.448731 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" podStartSLOduration=127.448704111 podStartE2EDuration="2m7.448704111s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:24.445438913 +0000 UTC m=+146.895154461" watchObservedRunningTime="2025-10-10 14:47:24.448704111 +0000 UTC m=+146.898419659" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.510919 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.516641 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.016623482 +0000 UTC m=+147.466339030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.584309 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.584823 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.614285 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.614494 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.114437389 +0000 UTC m=+147.564152937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.614724 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.615300 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.115274369 +0000 UTC m=+147.564989917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.715930 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.716217 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.216190891 +0000 UTC m=+147.665906439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.716590 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.716937 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.216925569 +0000 UTC m=+147.666641117 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.817906 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.818154 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.318100167 +0000 UTC m=+147.767815715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.818394 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.818858 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.318841205 +0000 UTC m=+147.768556753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.919954 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.920245 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.420206188 +0000 UTC m=+147.869921736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.920690 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:24 crc kubenswrapper[4788]: E1010 14:47:24.921067 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.421050798 +0000 UTC m=+147.870766346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:24 crc kubenswrapper[4788]: I1010 14:47:24.966631 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8txtm" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.021583 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.021885 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.521834087 +0000 UTC m=+147.971549635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.022226 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.022819 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.522793819 +0000 UTC m=+147.972509367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.063179 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.063264 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.123876 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.124331 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.624300416 +0000 UTC m=+148.074015964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.224958 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.225372 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.725356171 +0000 UTC m=+148.175071719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.295303 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" event={"ID":"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a","Type":"ContainerStarted","Data":"95d45dca47c61283fef4a79ffa78df851c6de47f8f641dff6f172630475ac1cd"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.296977 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" event={"ID":"f7897a5f-602b-4d6d-b266-4a80080d8671","Type":"ContainerStarted","Data":"be802456e23db910270fbaa13dd90759bbd74f7f9a7e0e5db8f11dbec84c26d1"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.297178 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.298754 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" event={"ID":"360b557a-fc75-4704-8cb2-fb9285d18566","Type":"ContainerStarted","Data":"5b64860ce6563245d25938e9d275352ca9642316229619b76eb8fe0b1ea3899d"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.299207 4788 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8tmcp container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.299321 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" podUID="f7897a5f-602b-4d6d-b266-4a80080d8671" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.300262 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" event={"ID":"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7","Type":"ContainerStarted","Data":"4ada8ee571c020d8f20aeaea434586929ac2dee8b96c474b1f09e1356223dbfd"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.300301 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" event={"ID":"ada58ad7-6a6a-459c-a313-e8c2d8e9a6e7","Type":"ContainerStarted","Data":"d8d29a48fc9571d0ef52eae07146c093f2ceb4d87611e752b7b116a43dee74a7"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.300514 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.305436 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" event={"ID":"c389c203-6864-4423-bb2a-da5693e052dc","Type":"ContainerStarted","Data":"2e13b99a80f0af99739a1a512bd896ecd5d7de64a5e4cb02a994ad2cd69b54ca"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.308115 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" event={"ID":"19cfebf6-d761-4be2-a146-082c2a71deba","Type":"ContainerStarted","Data":"06b416e9a94ea2c12dfa53262428ce39f1666b03b1f1a4fb4f9aebfc7e1065b8"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.308178 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" event={"ID":"19cfebf6-d761-4be2-a146-082c2a71deba","Type":"ContainerStarted","Data":"1d96d331f901501eef6d0e8a8a631f530e1ed42be56e28cb693f02c78eea60da"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.313775 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" event={"ID":"aad02b5d-b3d9-4005-90d4-8e824d28a95d","Type":"ContainerStarted","Data":"d17e13ff31697486cb9e5c3a3967714b394164416a5483966b5508c0348b6424"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.316058 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" event={"ID":"6b2c9335-c98e-4808-b9cf-ecac89cd1a99","Type":"ContainerStarted","Data":"422b7fdb84f1eadfa392b37ea1f60812ffa244d0d2f28167c55a6a15a12a9207"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.318218 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" event={"ID":"ffc68803-aa9e-4151-abc6-6d6a3eef7a8b","Type":"ContainerStarted","Data":"db7704370a92597957885e97718c0e940d905650466a7c5f1c4b77a22b77b459"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.325897 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" event={"ID":"6daf8c88-e9e8-48cf-acb8-c7a9f8cd9078","Type":"ContainerStarted","Data":"50e46e52c99c84a746fd057039d6df8f82087425dea4858472813847ca7301bf"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.326281 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.326408 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.826376726 +0000 UTC m=+148.276092274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.326724 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.327090 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.827073473 +0000 UTC m=+148.276789021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.330664 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" event={"ID":"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05","Type":"ContainerStarted","Data":"ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.330917 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.334466 4788 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-5j9sp container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" start-of-body= Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.334520 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" podUID="3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.336611 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" event={"ID":"05e27a21-9c12-471b-b959-0a8a2046a371","Type":"ContainerStarted","Data":"538c68040aa50952b966d6d91040e7a15e48888ea7bab9e4101d8cce714fbd10"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.336648 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" event={"ID":"05e27a21-9c12-471b-b959-0a8a2046a371","Type":"ContainerStarted","Data":"c1c2a9ba9c28646b5ab8adaf5ffb7af9e40fb5b88f2a4f1a2719b351b1734aa4"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.340799 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" event={"ID":"bb4a9174-7055-4a98-a3dc-790713348443","Type":"ContainerStarted","Data":"86de92268b664c77cd94e809cbcf7a96b53b70107a2f7960ecef968d0348817e"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.345099 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" event={"ID":"2f007d78-6f96-4523-8ae6-7d03955f0743","Type":"ContainerStarted","Data":"10403f00782b8304e2f2f71e531119b5598c2f8b0aa53f3fa22e0748bd73e42d"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.347782 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-jcksv" event={"ID":"cedd44c5-a11e-48e9-b61f-04c087777fc9","Type":"ContainerStarted","Data":"b352a11c76c0761ee318b6c7684e666cd421cafc2804c240d3d3bf0dc65c92e6"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.354380 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" event={"ID":"e7778a0c-9cb8-4c35-b5b0-7aa50a9290e7","Type":"ContainerStarted","Data":"89516f04234f4ff5f3c081aae083c57fe0a03001694d2502ffe06aeeef42d074"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.361205 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" event={"ID":"cc27468f-ca2c-4756-9d8f-d3b7a3325e3e","Type":"ContainerStarted","Data":"744142d9c32b131f5cb756a69ed3198cc9ca76e20e0bdd06698f18120b166aff"} Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.362980 4788 patch_prober.go:28] interesting pod/console-operator-58897d9998-cjsq9 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.363119 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" podUID="e7ec1502-208c-48bb-b8ad-8cb5ab938da2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.375843 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpvkv" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.379247 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lcsfw" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.379963 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:25 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:25 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:25 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.380013 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.428484 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.435501 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:25.935473804 +0000 UTC m=+148.385189352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.531700 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.532078 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.032064092 +0000 UTC m=+148.481779640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.567126 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" podStartSLOduration=127.567109923 podStartE2EDuration="2m7.567109923s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:25.505570386 +0000 UTC m=+147.955285934" watchObservedRunningTime="2025-10-10 14:47:25.567109923 +0000 UTC m=+148.016825471" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.634692 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.635073 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.135057395 +0000 UTC m=+148.584772943 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.647647 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.689178 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" podStartSLOduration=127.689154652 podStartE2EDuration="2m7.689154652s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:25.575797072 +0000 UTC m=+148.025512630" watchObservedRunningTime="2025-10-10 14:47:25.689154652 +0000 UTC m=+148.138870200" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.739649 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.740022 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.240009983 +0000 UTC m=+148.689725531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.760023 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-cgbfx" podStartSLOduration=127.760005933 podStartE2EDuration="2m7.760005933s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:25.759463239 +0000 UTC m=+148.209178787" watchObservedRunningTime="2025-10-10 14:47:25.760005933 +0000 UTC m=+148.209721481" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.760517 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-22mm5" podStartSLOduration=127.760512265 podStartE2EDuration="2m7.760512265s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:25.722554484 +0000 UTC m=+148.172270032" watchObservedRunningTime="2025-10-10 14:47:25.760512265 +0000 UTC m=+148.210227813" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.843888 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.844425 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.344401698 +0000 UTC m=+148.794117246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.854922 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" podStartSLOduration=127.854901991 podStartE2EDuration="2m7.854901991s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:25.851749344 +0000 UTC m=+148.301464892" watchObservedRunningTime="2025-10-10 14:47:25.854901991 +0000 UTC m=+148.304617539" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.886794 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hhq8v" podStartSLOduration=127.886762985 podStartE2EDuration="2m7.886762985s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:25.886479368 +0000 UTC m=+148.336194916" watchObservedRunningTime="2025-10-10 14:47:25.886762985 +0000 UTC m=+148.336478563" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.946833 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:25 crc kubenswrapper[4788]: E1010 14:47:25.947171 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.447136044 +0000 UTC m=+148.896851592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.988148 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2bt9s" podStartSLOduration=127.988110498 podStartE2EDuration="2m7.988110498s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:25.985412683 +0000 UTC m=+148.435128231" watchObservedRunningTime="2025-10-10 14:47:25.988110498 +0000 UTC m=+148.437826046" Oct 10 14:47:25 crc kubenswrapper[4788]: I1010 14:47:25.988986 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6ccmq" podStartSLOduration=127.988981168 podStartE2EDuration="2m7.988981168s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:25.927496852 +0000 UTC m=+148.377212400" watchObservedRunningTime="2025-10-10 14:47:25.988981168 +0000 UTC m=+148.438696716" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.028472 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6w9gx" podStartSLOduration=128.028454066 podStartE2EDuration="2m8.028454066s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:26.02072715 +0000 UTC m=+148.470442698" watchObservedRunningTime="2025-10-10 14:47:26.028454066 +0000 UTC m=+148.478169604" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.048242 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.048398 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.048441 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.048511 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.049739 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.549714926 +0000 UTC m=+148.999430474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.062271 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.062745 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.068030 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.069606 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qvvxg" podStartSLOduration=128.069589583 podStartE2EDuration="2m8.069589583s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:26.068390024 +0000 UTC m=+148.518105562" watchObservedRunningTime="2025-10-10 14:47:26.069589583 +0000 UTC m=+148.519305131" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.130944 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-459pb" podStartSLOduration=128.130924975 podStartE2EDuration="2m8.130924975s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:26.128504937 +0000 UTC m=+148.578220485" watchObservedRunningTime="2025-10-10 14:47:26.130924975 +0000 UTC m=+148.580640523" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.149638 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.149722 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.150053 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.650035293 +0000 UTC m=+149.099750841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.152757 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jkwgg"] Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.153162 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.154016 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.157594 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.185882 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-8vx6h" podStartSLOduration=128.185858783 podStartE2EDuration="2m8.185858783s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:26.174420639 +0000 UTC m=+148.624136187" watchObservedRunningTime="2025-10-10 14:47:26.185858783 +0000 UTC m=+148.635574331" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.271808 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.272997 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.273304 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-catalog-content\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.273413 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-utilities\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.273436 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fd57\" (UniqueName: \"kubernetes.io/projected/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-kube-api-access-7fd57\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.273552 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.773531168 +0000 UTC m=+149.223246716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.274860 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.275300 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.285824 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jkwgg"] Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.296461 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8dhqw" podStartSLOduration=128.296438557 podStartE2EDuration="2m8.296438557s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:26.273969038 +0000 UTC m=+148.723684596" watchObservedRunningTime="2025-10-10 14:47:26.296438557 +0000 UTC m=+148.746154155" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.383238 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-utilities\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.383295 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fd57\" (UniqueName: \"kubernetes.io/projected/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-kube-api-access-7fd57\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.383331 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-catalog-content\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.383383 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.383830 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.883812684 +0000 UTC m=+149.333528222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.384361 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-utilities\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.384983 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-catalog-content\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.385536 4788 patch_prober.go:28] interesting pod/apiserver-76f77b778f-2n42n container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]log ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]etcd ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/generic-apiserver-start-informers ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/max-in-flight-filter ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 10 14:47:26 crc kubenswrapper[4788]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 10 14:47:26 crc kubenswrapper[4788]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/project.openshift.io-projectcache ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 10 14:47:26 crc kubenswrapper[4788]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 10 14:47:26 crc kubenswrapper[4788]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 10 14:47:26 crc kubenswrapper[4788]: livez check failed Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.385599 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" podUID="fcf134f5-5c52-447f-b5b7-d3fbe2aef728" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.413466 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:26 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:26 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:26 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.413554 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.450299 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" event={"ID":"360b557a-fc75-4704-8cb2-fb9285d18566","Type":"ContainerStarted","Data":"a1fd43efe945d9715b41624228ed24fac72b1673154e1225cff34b074758fcc5"} Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.450337 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" event={"ID":"360b557a-fc75-4704-8cb2-fb9285d18566","Type":"ContainerStarted","Data":"62c34a52ea9a4b3e35ea28199a08732e6d4da176403538dd703f4eaa5bca93bc"} Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.472075 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-n2bnk" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.484312 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.487046 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:26.987021781 +0000 UTC m=+149.436737319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.505791 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fd57\" (UniqueName: \"kubernetes.io/projected/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-kube-api-access-7fd57\") pod \"certified-operators-jkwgg\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.540514 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.556629 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5b76v"] Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.578985 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.587883 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.588300 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.088288692 +0000 UTC m=+149.538004240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.618924 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5b76v"] Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.698673 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.698864 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-utilities\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.698922 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4zz8\" (UniqueName: \"kubernetes.io/projected/ff6f90f1-145e-41d4-9529-67d7929598a5-kube-api-access-d4zz8\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.698958 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-catalog-content\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.699107 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.199093272 +0000 UTC m=+149.648808820 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.780243 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6qwf8"] Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.781980 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.798059 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.805665 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-utilities\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.805745 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4zz8\" (UniqueName: \"kubernetes.io/projected/ff6f90f1-145e-41d4-9529-67d7929598a5-kube-api-access-d4zz8\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.805775 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.805794 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-catalog-content\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.806201 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-catalog-content\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.806456 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-utilities\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.806696 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.306685413 +0000 UTC m=+149.756400961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.906413 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.906674 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq4dz\" (UniqueName: \"kubernetes.io/projected/8308414f-3baf-4667-9b05-801f0fbde688-kube-api-access-fq4dz\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.906703 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-utilities\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.906760 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-catalog-content\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:26 crc kubenswrapper[4788]: E1010 14:47:26.906869 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.406853338 +0000 UTC m=+149.856568886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.918810 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4zz8\" (UniqueName: \"kubernetes.io/projected/ff6f90f1-145e-41d4-9529-67d7929598a5-kube-api-access-d4zz8\") pod \"certified-operators-5b76v\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.934309 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hggwc"] Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.935454 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.939361 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6qwf8"] Oct 10 14:47:26 crc kubenswrapper[4788]: I1010 14:47:26.956949 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hggwc"] Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.008665 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-catalog-content\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.008717 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq4dz\" (UniqueName: \"kubernetes.io/projected/8308414f-3baf-4667-9b05-801f0fbde688-kube-api-access-fq4dz\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.008746 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-utilities\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.008779 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.008802 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-catalog-content\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.008835 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-utilities\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.008851 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk2gv\" (UniqueName: \"kubernetes.io/projected/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-kube-api-access-dk2gv\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.009724 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-utilities\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.009948 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.509936772 +0000 UTC m=+149.959652320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.017305 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-catalog-content\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.050860 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq4dz\" (UniqueName: \"kubernetes.io/projected/8308414f-3baf-4667-9b05-801f0fbde688-kube-api-access-fq4dz\") pod \"community-operators-6qwf8\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.107403 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.111595 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.111884 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-utilities\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.111912 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk2gv\" (UniqueName: \"kubernetes.io/projected/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-kube-api-access-dk2gv\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.111953 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-catalog-content\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.112829 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-utilities\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.112851 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-catalog-content\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.112921 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.612905613 +0000 UTC m=+150.062621161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.168510 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-cjsq9" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.195887 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk2gv\" (UniqueName: \"kubernetes.io/projected/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-kube-api-access-dk2gv\") pod \"community-operators-hggwc\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.203054 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.213423 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.213816 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.713803814 +0000 UTC m=+150.163519362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.274938 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.317698 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.318010 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.817995476 +0000 UTC m=+150.267711024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.388870 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:27 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:27 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:27 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.388926 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.426031 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.426876 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:27.926863738 +0000 UTC m=+150.376579286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.453202 4788 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-5j9sp container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.24:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.453253 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" podUID="3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.24:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.453485 4788 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8tmcp container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.453503 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" podUID="f7897a5f-602b-4d6d-b266-4a80080d8671" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.528403 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" event={"ID":"360b557a-fc75-4704-8cb2-fb9285d18566","Type":"ContainerStarted","Data":"ce92ff7b2be01c3e1f9ee2da30ead610e509b76135001bbed83ff0e0c511c670"} Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.533870 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.534182 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.034166063 +0000 UTC m=+150.483881611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.542927 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ac486cd6bbd117152020d928b3c2ed2d3e709d576b6fe8e9166cc22e5a43efc5"} Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.573972 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-xfl9r" podStartSLOduration=10.573955978 podStartE2EDuration="10.573955978s" podCreationTimestamp="2025-10-10 14:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:27.573091068 +0000 UTC m=+150.022806616" watchObservedRunningTime="2025-10-10 14:47:27.573955978 +0000 UTC m=+150.023671526" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.638906 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.641110 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.141098069 +0000 UTC m=+150.590813607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.721490 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jkwgg"] Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.741868 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.742239 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.242205167 +0000 UTC m=+150.691920715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: W1010 14:47:27.800464 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-10c424e1b760443d7649a3e95126b05220e949bc2414e74b8a3e6fd169bc1f9e WatchSource:0}: Error finding container 10c424e1b760443d7649a3e95126b05220e949bc2414e74b8a3e6fd169bc1f9e: Status 404 returned error can't find the container with id 10c424e1b760443d7649a3e95126b05220e949bc2414e74b8a3e6fd169bc1f9e Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.821478 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6qwf8"] Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.843870 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.844202 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.344191373 +0000 UTC m=+150.793906921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:27 crc kubenswrapper[4788]: W1010 14:47:27.862321 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8308414f_3baf_4667_9b05_801f0fbde688.slice/crio-20ab22c578d6d1575a10d3f78e2b942546c037637ea8e18178969896e6523ab3 WatchSource:0}: Error finding container 20ab22c578d6d1575a10d3f78e2b942546c037637ea8e18178969896e6523ab3: Status 404 returned error can't find the container with id 20ab22c578d6d1575a10d3f78e2b942546c037637ea8e18178969896e6523ab3 Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.871638 4788 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 10 14:47:27 crc kubenswrapper[4788]: I1010 14:47:27.951690 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:27 crc kubenswrapper[4788]: E1010 14:47:27.952125 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.452107113 +0000 UTC m=+150.901822661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.021270 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hggwc"] Oct 10 14:47:28 crc kubenswrapper[4788]: W1010 14:47:28.046963 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9b85b1a_ac5c_4aad_8183_dcd0fb4cf5c6.slice/crio-ccc28281296d8f8567b8b72a8aa273fe8b3499e6a5404fe8006ea0806b8a3a56 WatchSource:0}: Error finding container ccc28281296d8f8567b8b72a8aa273fe8b3499e6a5404fe8006ea0806b8a3a56: Status 404 returned error can't find the container with id ccc28281296d8f8567b8b72a8aa273fe8b3499e6a5404fe8006ea0806b8a3a56 Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.053268 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:28 crc kubenswrapper[4788]: E1010 14:47:28.053695 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.553674901 +0000 UTC m=+151.003390449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.107963 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5b76v"] Oct 10 14:47:28 crc kubenswrapper[4788]: W1010 14:47:28.119474 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff6f90f1_145e_41d4_9529_67d7929598a5.slice/crio-1f4b4572e8c0898a357dc4b0a33444244aeb44b2011b1898a59efc320127b788 WatchSource:0}: Error finding container 1f4b4572e8c0898a357dc4b0a33444244aeb44b2011b1898a59efc320127b788: Status 404 returned error can't find the container with id 1f4b4572e8c0898a357dc4b0a33444244aeb44b2011b1898a59efc320127b788 Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.153977 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:28 crc kubenswrapper[4788]: E1010 14:47:28.154245 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.654204514 +0000 UTC m=+151.103920062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.154330 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:28 crc kubenswrapper[4788]: E1010 14:47:28.154966 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.654950552 +0000 UTC m=+151.104666090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.255912 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:28 crc kubenswrapper[4788]: E1010 14:47:28.256104 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.756070209 +0000 UTC m=+151.205785757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.256355 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:28 crc kubenswrapper[4788]: E1010 14:47:28.256947 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.756921589 +0000 UTC m=+151.206637137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.315212 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mkxhd"] Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.316659 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.318875 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.324841 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkxhd"] Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.357825 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:28 crc kubenswrapper[4788]: E1010 14:47:28.358073 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.858039006 +0000 UTC m=+151.307754554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.368307 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv4nr\" (UniqueName: \"kubernetes.io/projected/10662620-8659-4eb6-b84d-420c73f200bb-kube-api-access-dv4nr\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.368505 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-utilities\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.368611 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.370073 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-catalog-content\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: E1010 14:47:28.371317 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-10 14:47:28.871286424 +0000 UTC m=+151.321001972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kdbvl" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.379697 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:28 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:28 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:28 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.379763 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.435906 4788 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-10T14:47:27.871663653Z","Handler":null,"Name":""} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.439490 4788 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.439535 4788 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.472088 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.472784 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-utilities\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.472942 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-catalog-content\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.473042 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv4nr\" (UniqueName: \"kubernetes.io/projected/10662620-8659-4eb6-b84d-420c73f200bb-kube-api-access-dv4nr\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.474703 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-utilities\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.475014 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-catalog-content\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.490404 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.515178 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv4nr\" (UniqueName: \"kubernetes.io/projected/10662620-8659-4eb6-b84d-420c73f200bb-kube-api-access-dv4nr\") pod \"redhat-marketplace-mkxhd\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.554184 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e1a40092300f65c5f4cc80d70f3d1997a420b48dd502c808fea1c189b3117b79"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.554248 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9d6603b149fba31421d229408dfe954fd1a40ee1e1e7c10e81c53ac0db37646b"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.557022 4788 generic.go:334] "Generic (PLEG): container finished" podID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerID="157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8" exitCode=0 Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.557502 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b76v" event={"ID":"ff6f90f1-145e-41d4-9529-67d7929598a5","Type":"ContainerDied","Data":"157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.557540 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b76v" event={"ID":"ff6f90f1-145e-41d4-9529-67d7929598a5","Type":"ContainerStarted","Data":"1f4b4572e8c0898a357dc4b0a33444244aeb44b2011b1898a59efc320127b788"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.560625 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.574236 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e5212c16c7b78f8e71e0710d57456afb634cb6be0fb669e7a12d2069e0e48ac9"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.574319 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"10c424e1b760443d7649a3e95126b05220e949bc2414e74b8a3e6fd169bc1f9e"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.574480 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.575292 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.586590 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerID="df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579" exitCode=0 Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.586739 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hggwc" event={"ID":"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6","Type":"ContainerDied","Data":"df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.586773 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hggwc" event={"ID":"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6","Type":"ContainerStarted","Data":"ccc28281296d8f8567b8b72a8aa273fe8b3499e6a5404fe8006ea0806b8a3a56"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.587152 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.587202 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.590826 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c3caa44f30ea6d2958bd39efc37db24c7ead686c836e626b2e4d9dc9f62ef2ba"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.596849 4788 generic.go:334] "Generic (PLEG): container finished" podID="8308414f-3baf-4667-9b05-801f0fbde688" containerID="dc73e01b5493baef27fda9bfd6257d4b95bca0665cf39b2aa830a0ed6236385a" exitCode=0 Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.596943 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qwf8" event={"ID":"8308414f-3baf-4667-9b05-801f0fbde688","Type":"ContainerDied","Data":"dc73e01b5493baef27fda9bfd6257d4b95bca0665cf39b2aa830a0ed6236385a"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.597013 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qwf8" event={"ID":"8308414f-3baf-4667-9b05-801f0fbde688","Type":"ContainerStarted","Data":"20ab22c578d6d1575a10d3f78e2b942546c037637ea8e18178969896e6523ab3"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.600057 4788 generic.go:334] "Generic (PLEG): container finished" podID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerID="08066560a1488d814ab12d68ba5b0d935e8aa03a30e435c6570b213239ef8840" exitCode=0 Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.600186 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkwgg" event={"ID":"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100","Type":"ContainerDied","Data":"08066560a1488d814ab12d68ba5b0d935e8aa03a30e435c6570b213239ef8840"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.600241 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkwgg" event={"ID":"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100","Type":"ContainerStarted","Data":"0fc4c9f421e9214a63b17095e0597c011e0b7967e18cc98a9170465c68f1d8bf"} Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.619777 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kdbvl\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.703589 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t7brq"] Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.706590 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.712483 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7brq"] Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.717335 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.777113 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-catalog-content\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.777199 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxzc4\" (UniqueName: \"kubernetes.io/projected/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-kube-api-access-hxzc4\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.777223 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-utilities\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.798700 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.878979 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-catalog-content\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.879367 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxzc4\" (UniqueName: \"kubernetes.io/projected/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-kube-api-access-hxzc4\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.879386 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-utilities\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.879936 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-utilities\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.880156 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-catalog-content\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.904852 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxzc4\" (UniqueName: \"kubernetes.io/projected/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-kube-api-access-hxzc4\") pod \"redhat-marketplace-t7brq\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:28 crc kubenswrapper[4788]: I1010 14:47:28.959741 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkxhd"] Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.031686 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.153486 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kdbvl"] Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.342999 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7brq"] Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.377262 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:29 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:29 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:29 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.377583 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.406122 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.406654 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.502121 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-662r8"] Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.503184 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.505519 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.521292 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-662r8"] Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.589848 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqkm5\" (UniqueName: \"kubernetes.io/projected/9479f6c6-ef3d-4a69-affe-2f773e32b48f-kube-api-access-jqkm5\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.590019 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-catalog-content\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.590053 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-utilities\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.595438 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.606790 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2n42n" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.646061 4788 generic.go:334] "Generic (PLEG): container finished" podID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerID="2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36" exitCode=0 Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.646469 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7brq" event={"ID":"8ef8ef60-95ef-4e3d-8e13-149f602fc08d","Type":"ContainerDied","Data":"2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36"} Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.646507 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7brq" event={"ID":"8ef8ef60-95ef-4e3d-8e13-149f602fc08d","Type":"ContainerStarted","Data":"43c95117a18a896c5e7087075f3c71c3955b25fff50d850e6a93e2bc8e726e17"} Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.649888 4788 generic.go:334] "Generic (PLEG): container finished" podID="10662620-8659-4eb6-b84d-420c73f200bb" containerID="939fe39905084927723b11c348b2e4f8d197212ac85f9c86c776a179b4f8f48e" exitCode=0 Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.650019 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkxhd" event={"ID":"10662620-8659-4eb6-b84d-420c73f200bb","Type":"ContainerDied","Data":"939fe39905084927723b11c348b2e4f8d197212ac85f9c86c776a179b4f8f48e"} Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.650123 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkxhd" event={"ID":"10662620-8659-4eb6-b84d-420c73f200bb","Type":"ContainerStarted","Data":"26bbdb6a3a54d23985a644fe7b256b5cce67186558ceaf0952a5f611836129ec"} Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.690306 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" event={"ID":"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce","Type":"ContainerStarted","Data":"fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8"} Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.690359 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" event={"ID":"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce","Type":"ContainerStarted","Data":"48e63761b0c12c77c6717431c9a535c4b6529012d464fc08970afe59064efcb3"} Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.690645 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.690742 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqkm5\" (UniqueName: \"kubernetes.io/projected/9479f6c6-ef3d-4a69-affe-2f773e32b48f-kube-api-access-jqkm5\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.690937 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-catalog-content\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.690963 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-utilities\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.692265 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-catalog-content\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.692556 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-utilities\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.738294 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqkm5\" (UniqueName: \"kubernetes.io/projected/9479f6c6-ef3d-4a69-affe-2f773e32b48f-kube-api-access-jqkm5\") pod \"redhat-operators-662r8\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.789995 4788 patch_prober.go:28] interesting pod/downloads-7954f5f757-h28k6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.790041 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-h28k6" podUID="7d2c7d67-2055-4e2b-9598-4438c3b702c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.790055 4788 patch_prober.go:28] interesting pod/downloads-7954f5f757-h28k6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.790110 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h28k6" podUID="7d2c7d67-2055-4e2b-9598-4438c3b702c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.812646 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" podStartSLOduration=131.812629226 podStartE2EDuration="2m11.812629226s" podCreationTimestamp="2025-10-10 14:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:29.809831379 +0000 UTC m=+152.259546927" watchObservedRunningTime="2025-10-10 14:47:29.812629226 +0000 UTC m=+152.262344774" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.881825 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.919267 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vzskc"] Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.920449 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.963447 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vzskc"] Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.997562 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wn2f\" (UniqueName: \"kubernetes.io/projected/5bec33de-d109-4518-9a02-a9a65382a53b-kube-api-access-9wn2f\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.997663 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-utilities\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:29 crc kubenswrapper[4788]: I1010 14:47:29.997686 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-catalog-content\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.101855 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-utilities\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.102286 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-catalog-content\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.102416 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wn2f\" (UniqueName: \"kubernetes.io/projected/5bec33de-d109-4518-9a02-a9a65382a53b-kube-api-access-9wn2f\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.103677 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-catalog-content\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.104022 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-utilities\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.130425 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wn2f\" (UniqueName: \"kubernetes.io/projected/5bec33de-d109-4518-9a02-a9a65382a53b-kube-api-access-9wn2f\") pod \"redhat-operators-vzskc\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.161247 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.161328 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.165180 4788 patch_prober.go:28] interesting pod/console-f9d7485db-mgbhm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.165249 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mgbhm" podUID="d8826b11-1579-4653-953e-0895bc1d16b5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.197213 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.241631 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.267647 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.268225 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.277525 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.277730 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.284613 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-662r8"] Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.294352 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.294626 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.309087 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8612809d-f9ea-4c98-9eea-26f9540d8613-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8612809d-f9ea-4c98-9eea-26f9540d8613\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.309884 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8612809d-f9ea-4c98-9eea-26f9540d8613-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8612809d-f9ea-4c98-9eea-26f9540d8613\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:30 crc kubenswrapper[4788]: W1010 14:47:30.314863 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9479f6c6_ef3d_4a69_affe_2f773e32b48f.slice/crio-1fd4aec502fc6e0a085d1e29e394ed137469eb183d2949dae8d0928b32e13dba WatchSource:0}: Error finding container 1fd4aec502fc6e0a085d1e29e394ed137469eb183d2949dae8d0928b32e13dba: Status 404 returned error can't find the container with id 1fd4aec502fc6e0a085d1e29e394ed137469eb183d2949dae8d0928b32e13dba Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.374466 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.390859 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:30 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:30 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:30 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.391314 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.413015 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8612809d-f9ea-4c98-9eea-26f9540d8613-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8612809d-f9ea-4c98-9eea-26f9540d8613\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.413173 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8612809d-f9ea-4c98-9eea-26f9540d8613-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8612809d-f9ea-4c98-9eea-26f9540d8613\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.415344 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8612809d-f9ea-4c98-9eea-26f9540d8613-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8612809d-f9ea-4c98-9eea-26f9540d8613\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.438985 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8612809d-f9ea-4c98-9eea-26f9540d8613-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8612809d-f9ea-4c98-9eea-26f9540d8613\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.610205 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8tmcp" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.659406 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.717124 4788 generic.go:334] "Generic (PLEG): container finished" podID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerID="2dec899db81718eb3d070f1d2a3cbed752f3853591457578060270a11237b541" exitCode=0 Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.719372 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-662r8" event={"ID":"9479f6c6-ef3d-4a69-affe-2f773e32b48f","Type":"ContainerDied","Data":"2dec899db81718eb3d070f1d2a3cbed752f3853591457578060270a11237b541"} Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.719420 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-662r8" event={"ID":"9479f6c6-ef3d-4a69-affe-2f773e32b48f","Type":"ContainerStarted","Data":"1fd4aec502fc6e0a085d1e29e394ed137469eb183d2949dae8d0928b32e13dba"} Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.747483 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:47:30 crc kubenswrapper[4788]: I1010 14:47:30.820072 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vzskc"] Oct 10 14:47:31 crc kubenswrapper[4788]: I1010 14:47:31.109997 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 10 14:47:31 crc kubenswrapper[4788]: I1010 14:47:31.404475 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:31 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:31 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:31 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:31 crc kubenswrapper[4788]: I1010 14:47:31.405214 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:31 crc kubenswrapper[4788]: I1010 14:47:31.727628 4788 generic.go:334] "Generic (PLEG): container finished" podID="5bec33de-d109-4518-9a02-a9a65382a53b" containerID="64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e" exitCode=0 Oct 10 14:47:31 crc kubenswrapper[4788]: I1010 14:47:31.727718 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzskc" event={"ID":"5bec33de-d109-4518-9a02-a9a65382a53b","Type":"ContainerDied","Data":"64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e"} Oct 10 14:47:31 crc kubenswrapper[4788]: I1010 14:47:31.727788 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzskc" event={"ID":"5bec33de-d109-4518-9a02-a9a65382a53b","Type":"ContainerStarted","Data":"55a86c4757aa6848f91690aa4007a69397a1961e1621a5d09221fff39937fecc"} Oct 10 14:47:31 crc kubenswrapper[4788]: I1010 14:47:31.735806 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8612809d-f9ea-4c98-9eea-26f9540d8613","Type":"ContainerStarted","Data":"1fe1a700804fc37fc27dd6dd3ad315e711c45550e7dd2cca11d6b10ca2b30a55"} Oct 10 14:47:32 crc kubenswrapper[4788]: I1010 14:47:32.063939 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6xm66" Oct 10 14:47:32 crc kubenswrapper[4788]: I1010 14:47:32.378679 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:32 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:32 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:32 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:32 crc kubenswrapper[4788]: I1010 14:47:32.378744 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:32 crc kubenswrapper[4788]: I1010 14:47:32.752622 4788 generic.go:334] "Generic (PLEG): container finished" podID="2bf9df01-bc0d-46cf-8786-0ab3acea7a7a" containerID="95d45dca47c61283fef4a79ffa78df851c6de47f8f641dff6f172630475ac1cd" exitCode=0 Oct 10 14:47:32 crc kubenswrapper[4788]: I1010 14:47:32.752717 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" event={"ID":"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a","Type":"ContainerDied","Data":"95d45dca47c61283fef4a79ffa78df851c6de47f8f641dff6f172630475ac1cd"} Oct 10 14:47:32 crc kubenswrapper[4788]: I1010 14:47:32.776885 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8612809d-f9ea-4c98-9eea-26f9540d8613","Type":"ContainerStarted","Data":"45c2959211c5ac379f0b3c72e88264cd74b9091345ff6dbca30a4568796b2bd3"} Oct 10 14:47:32 crc kubenswrapper[4788]: I1010 14:47:32.797417 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.797391422 podStartE2EDuration="2.797391422s" podCreationTimestamp="2025-10-10 14:47:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:32.794094693 +0000 UTC m=+155.243810241" watchObservedRunningTime="2025-10-10 14:47:32.797391422 +0000 UTC m=+155.247106970" Oct 10 14:47:33 crc kubenswrapper[4788]: I1010 14:47:33.376961 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:33 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:33 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:33 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:33 crc kubenswrapper[4788]: I1010 14:47:33.377062 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:33 crc kubenswrapper[4788]: I1010 14:47:33.799734 4788 generic.go:334] "Generic (PLEG): container finished" podID="8612809d-f9ea-4c98-9eea-26f9540d8613" containerID="45c2959211c5ac379f0b3c72e88264cd74b9091345ff6dbca30a4568796b2bd3" exitCode=0 Oct 10 14:47:33 crc kubenswrapper[4788]: I1010 14:47:33.799837 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8612809d-f9ea-4c98-9eea-26f9540d8613","Type":"ContainerDied","Data":"45c2959211c5ac379f0b3c72e88264cd74b9091345ff6dbca30a4568796b2bd3"} Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.161341 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.182175 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 14:47:34 crc kubenswrapper[4788]: E1010 14:47:34.182436 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf9df01-bc0d-46cf-8786-0ab3acea7a7a" containerName="collect-profiles" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.182455 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf9df01-bc0d-46cf-8786-0ab3acea7a7a" containerName="collect-profiles" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.182583 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf9df01-bc0d-46cf-8786-0ab3acea7a7a" containerName="collect-profiles" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.182959 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.196056 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.196057 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.208668 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.228347 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-config-volume\") pod \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.228575 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpwwz\" (UniqueName: \"kubernetes.io/projected/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-kube-api-access-jpwwz\") pod \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.228603 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-secret-volume\") pod \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\" (UID: \"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a\") " Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.228768 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a630608-9824-4339-a050-8c1c83b88b5b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7a630608-9824-4339-a050-8c1c83b88b5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.228860 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a630608-9824-4339-a050-8c1c83b88b5b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7a630608-9824-4339-a050-8c1c83b88b5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.229352 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-config-volume" (OuterVolumeSpecName: "config-volume") pod "2bf9df01-bc0d-46cf-8786-0ab3acea7a7a" (UID: "2bf9df01-bc0d-46cf-8786-0ab3acea7a7a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.247814 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-kube-api-access-jpwwz" (OuterVolumeSpecName: "kube-api-access-jpwwz") pod "2bf9df01-bc0d-46cf-8786-0ab3acea7a7a" (UID: "2bf9df01-bc0d-46cf-8786-0ab3acea7a7a"). InnerVolumeSpecName "kube-api-access-jpwwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.249813 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2bf9df01-bc0d-46cf-8786-0ab3acea7a7a" (UID: "2bf9df01-bc0d-46cf-8786-0ab3acea7a7a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.329767 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a630608-9824-4339-a050-8c1c83b88b5b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7a630608-9824-4339-a050-8c1c83b88b5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.329837 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a630608-9824-4339-a050-8c1c83b88b5b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7a630608-9824-4339-a050-8c1c83b88b5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.329880 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a630608-9824-4339-a050-8c1c83b88b5b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7a630608-9824-4339-a050-8c1c83b88b5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.329903 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpwwz\" (UniqueName: \"kubernetes.io/projected/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-kube-api-access-jpwwz\") on node \"crc\" DevicePath \"\"" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.329989 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.330013 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.347651 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a630608-9824-4339-a050-8c1c83b88b5b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7a630608-9824-4339-a050-8c1c83b88b5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.376375 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:34 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:34 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:34 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.376445 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.529058 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.864480 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" event={"ID":"2bf9df01-bc0d-46cf-8786-0ab3acea7a7a","Type":"ContainerDied","Data":"1f325e4b115eab39402ec0157557788fe7f0bcddab951f36d250ed67a0fb4dc3"} Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.864893 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f325e4b115eab39402ec0157557788fe7f0bcddab951f36d250ed67a0fb4dc3" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.864794 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r" Oct 10 14:47:34 crc kubenswrapper[4788]: I1010 14:47:34.885840 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 10 14:47:34 crc kubenswrapper[4788]: W1010 14:47:34.904330 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7a630608_9824_4339_a050_8c1c83b88b5b.slice/crio-eae617519c735625be7eb6d5b14fa2f8b39663f5da21584e0787a54c3ed95740 WatchSource:0}: Error finding container eae617519c735625be7eb6d5b14fa2f8b39663f5da21584e0787a54c3ed95740: Status 404 returned error can't find the container with id eae617519c735625be7eb6d5b14fa2f8b39663f5da21584e0787a54c3ed95740 Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.378282 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:35 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:35 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:35 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.378831 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.418594 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.468497 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8612809d-f9ea-4c98-9eea-26f9540d8613-kube-api-access\") pod \"8612809d-f9ea-4c98-9eea-26f9540d8613\" (UID: \"8612809d-f9ea-4c98-9eea-26f9540d8613\") " Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.468585 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8612809d-f9ea-4c98-9eea-26f9540d8613-kubelet-dir\") pod \"8612809d-f9ea-4c98-9eea-26f9540d8613\" (UID: \"8612809d-f9ea-4c98-9eea-26f9540d8613\") " Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.469116 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8612809d-f9ea-4c98-9eea-26f9540d8613-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8612809d-f9ea-4c98-9eea-26f9540d8613" (UID: "8612809d-f9ea-4c98-9eea-26f9540d8613"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.479906 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8612809d-f9ea-4c98-9eea-26f9540d8613-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8612809d-f9ea-4c98-9eea-26f9540d8613" (UID: "8612809d-f9ea-4c98-9eea-26f9540d8613"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.570669 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8612809d-f9ea-4c98-9eea-26f9540d8613-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.571094 4788 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8612809d-f9ea-4c98-9eea-26f9540d8613-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.913774 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7a630608-9824-4339-a050-8c1c83b88b5b","Type":"ContainerStarted","Data":"13842439f52f21eaaca644a885d1d30a782791181844dd6174107628fb403e98"} Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.913851 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7a630608-9824-4339-a050-8c1c83b88b5b","Type":"ContainerStarted","Data":"eae617519c735625be7eb6d5b14fa2f8b39663f5da21584e0787a54c3ed95740"} Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.938756 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.9387378750000002 podStartE2EDuration="1.938737875s" podCreationTimestamp="2025-10-10 14:47:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:35.936882209 +0000 UTC m=+158.386597747" watchObservedRunningTime="2025-10-10 14:47:35.938737875 +0000 UTC m=+158.388453423" Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.970040 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8612809d-f9ea-4c98-9eea-26f9540d8613","Type":"ContainerDied","Data":"1fe1a700804fc37fc27dd6dd3ad315e711c45550e7dd2cca11d6b10ca2b30a55"} Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.970078 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fe1a700804fc37fc27dd6dd3ad315e711c45550e7dd2cca11d6b10ca2b30a55" Oct 10 14:47:35 crc kubenswrapper[4788]: I1010 14:47:35.970185 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 10 14:47:36 crc kubenswrapper[4788]: I1010 14:47:36.376936 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:36 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:36 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:36 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:36 crc kubenswrapper[4788]: I1010 14:47:36.376993 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:36 crc kubenswrapper[4788]: I1010 14:47:36.993285 4788 generic.go:334] "Generic (PLEG): container finished" podID="7a630608-9824-4339-a050-8c1c83b88b5b" containerID="13842439f52f21eaaca644a885d1d30a782791181844dd6174107628fb403e98" exitCode=0 Oct 10 14:47:36 crc kubenswrapper[4788]: I1010 14:47:36.993410 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7a630608-9824-4339-a050-8c1c83b88b5b","Type":"ContainerDied","Data":"13842439f52f21eaaca644a885d1d30a782791181844dd6174107628fb403e98"} Oct 10 14:47:37 crc kubenswrapper[4788]: I1010 14:47:37.376201 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:37 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:37 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:37 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:37 crc kubenswrapper[4788]: I1010 14:47:37.376321 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:38 crc kubenswrapper[4788]: I1010 14:47:38.377262 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:38 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:38 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:38 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:38 crc kubenswrapper[4788]: I1010 14:47:38.377762 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:39 crc kubenswrapper[4788]: I1010 14:47:39.378404 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:39 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:39 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:39 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:39 crc kubenswrapper[4788]: I1010 14:47:39.378895 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:39 crc kubenswrapper[4788]: I1010 14:47:39.802642 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-h28k6" Oct 10 14:47:39 crc kubenswrapper[4788]: I1010 14:47:39.858083 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:39 crc kubenswrapper[4788]: I1010 14:47:39.885817 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3d61e9aa-0688-436b-a8af-d6d8353c4350-metrics-certs\") pod \"network-metrics-daemon-jqlwd\" (UID: \"3d61e9aa-0688-436b-a8af-d6d8353c4350\") " pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:40 crc kubenswrapper[4788]: I1010 14:47:40.049477 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jqlwd" Oct 10 14:47:40 crc kubenswrapper[4788]: I1010 14:47:40.161424 4788 patch_prober.go:28] interesting pod/console-f9d7485db-mgbhm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 10 14:47:40 crc kubenswrapper[4788]: I1010 14:47:40.161527 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mgbhm" podUID="d8826b11-1579-4653-953e-0895bc1d16b5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 10 14:47:40 crc kubenswrapper[4788]: I1010 14:47:40.376387 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:40 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:40 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:40 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:40 crc kubenswrapper[4788]: I1010 14:47:40.376446 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:41 crc kubenswrapper[4788]: I1010 14:47:41.377933 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:41 crc kubenswrapper[4788]: [-]has-synced failed: reason withheld Oct 10 14:47:41 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:41 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:41 crc kubenswrapper[4788]: I1010 14:47:41.378032 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:42 crc kubenswrapper[4788]: I1010 14:47:42.377688 4788 patch_prober.go:28] interesting pod/router-default-5444994796-99fw8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 10 14:47:42 crc kubenswrapper[4788]: [+]has-synced ok Oct 10 14:47:42 crc kubenswrapper[4788]: [+]process-running ok Oct 10 14:47:42 crc kubenswrapper[4788]: healthz check failed Oct 10 14:47:42 crc kubenswrapper[4788]: I1010 14:47:42.378076 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-99fw8" podUID="1ce4a5b7-557b-4b26-924c-2e246030b180" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 10 14:47:43 crc kubenswrapper[4788]: I1010 14:47:43.376574 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:43 crc kubenswrapper[4788]: I1010 14:47:43.379827 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-99fw8" Oct 10 14:47:44 crc kubenswrapper[4788]: I1010 14:47:44.414077 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:44 crc kubenswrapper[4788]: I1010 14:47:44.540634 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a630608-9824-4339-a050-8c1c83b88b5b-kubelet-dir\") pod \"7a630608-9824-4339-a050-8c1c83b88b5b\" (UID: \"7a630608-9824-4339-a050-8c1c83b88b5b\") " Oct 10 14:47:44 crc kubenswrapper[4788]: I1010 14:47:44.540688 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a630608-9824-4339-a050-8c1c83b88b5b-kube-api-access\") pod \"7a630608-9824-4339-a050-8c1c83b88b5b\" (UID: \"7a630608-9824-4339-a050-8c1c83b88b5b\") " Oct 10 14:47:44 crc kubenswrapper[4788]: I1010 14:47:44.540743 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a630608-9824-4339-a050-8c1c83b88b5b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7a630608-9824-4339-a050-8c1c83b88b5b" (UID: "7a630608-9824-4339-a050-8c1c83b88b5b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:47:44 crc kubenswrapper[4788]: I1010 14:47:44.541017 4788 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a630608-9824-4339-a050-8c1c83b88b5b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 10 14:47:44 crc kubenswrapper[4788]: I1010 14:47:44.547299 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a630608-9824-4339-a050-8c1c83b88b5b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7a630608-9824-4339-a050-8c1c83b88b5b" (UID: "7a630608-9824-4339-a050-8c1c83b88b5b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:47:44 crc kubenswrapper[4788]: I1010 14:47:44.641920 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a630608-9824-4339-a050-8c1c83b88b5b-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 10 14:47:45 crc kubenswrapper[4788]: I1010 14:47:45.080022 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7a630608-9824-4339-a050-8c1c83b88b5b","Type":"ContainerDied","Data":"eae617519c735625be7eb6d5b14fa2f8b39663f5da21584e0787a54c3ed95740"} Oct 10 14:47:45 crc kubenswrapper[4788]: I1010 14:47:45.080523 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eae617519c735625be7eb6d5b14fa2f8b39663f5da21584e0787a54c3ed95740" Oct 10 14:47:45 crc kubenswrapper[4788]: I1010 14:47:45.080297 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 10 14:47:46 crc kubenswrapper[4788]: I1010 14:47:46.263635 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jqlwd"] Oct 10 14:47:46 crc kubenswrapper[4788]: W1010 14:47:46.281932 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d61e9aa_0688_436b_a8af_d6d8353c4350.slice/crio-54a442c30f007f1322af64193359b6d30f50c63242f576b28406b29ee6909665 WatchSource:0}: Error finding container 54a442c30f007f1322af64193359b6d30f50c63242f576b28406b29ee6909665: Status 404 returned error can't find the container with id 54a442c30f007f1322af64193359b6d30f50c63242f576b28406b29ee6909665 Oct 10 14:47:47 crc kubenswrapper[4788]: I1010 14:47:47.095059 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" event={"ID":"3d61e9aa-0688-436b-a8af-d6d8353c4350","Type":"ContainerStarted","Data":"54a442c30f007f1322af64193359b6d30f50c63242f576b28406b29ee6909665"} Oct 10 14:47:48 crc kubenswrapper[4788]: I1010 14:47:48.102167 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" event={"ID":"3d61e9aa-0688-436b-a8af-d6d8353c4350","Type":"ContainerStarted","Data":"c3d78a5e1482da6c66abdcdbb1d3b940c4c2cace615ee271970c81ec5b71899e"} Oct 10 14:47:48 crc kubenswrapper[4788]: I1010 14:47:48.806149 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:47:50 crc kubenswrapper[4788]: I1010 14:47:50.169188 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:50 crc kubenswrapper[4788]: I1010 14:47:50.176593 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:47:58 crc kubenswrapper[4788]: I1010 14:47:58.192010 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jqlwd" event={"ID":"3d61e9aa-0688-436b-a8af-d6d8353c4350","Type":"ContainerStarted","Data":"72fbe8cff531153ac7cca1ba93a07e1c99e044390c9c9d5cef8aeca5d6c11651"} Oct 10 14:47:58 crc kubenswrapper[4788]: I1010 14:47:58.200288 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkwgg" event={"ID":"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100","Type":"ContainerStarted","Data":"5fe194d114f06da7c83ea5e21969f4eec3793d43a44602612ea8643a5115ab3b"} Oct 10 14:47:58 crc kubenswrapper[4788]: I1010 14:47:58.205096 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkxhd" event={"ID":"10662620-8659-4eb6-b84d-420c73f200bb","Type":"ContainerStarted","Data":"96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a"} Oct 10 14:47:58 crc kubenswrapper[4788]: I1010 14:47:58.213054 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-jqlwd" podStartSLOduration=161.213027153 podStartE2EDuration="2m41.213027153s" podCreationTimestamp="2025-10-10 14:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:47:58.210993084 +0000 UTC m=+180.660708672" watchObservedRunningTime="2025-10-10 14:47:58.213027153 +0000 UTC m=+180.662742711" Oct 10 14:47:58 crc kubenswrapper[4788]: I1010 14:47:58.231271 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-662r8" event={"ID":"9479f6c6-ef3d-4a69-affe-2f773e32b48f","Type":"ContainerStarted","Data":"23a5a08ccea022a599f9335fe459f5892e9afef159120ca632f864dc9df311a1"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.243000 4788 generic.go:334] "Generic (PLEG): container finished" podID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerID="23a5a08ccea022a599f9335fe459f5892e9afef159120ca632f864dc9df311a1" exitCode=0 Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.243092 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-662r8" event={"ID":"9479f6c6-ef3d-4a69-affe-2f773e32b48f","Type":"ContainerDied","Data":"23a5a08ccea022a599f9335fe459f5892e9afef159120ca632f864dc9df311a1"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.248411 4788 generic.go:334] "Generic (PLEG): container finished" podID="5bec33de-d109-4518-9a02-a9a65382a53b" containerID="010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746" exitCode=0 Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.248624 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzskc" event={"ID":"5bec33de-d109-4518-9a02-a9a65382a53b","Type":"ContainerDied","Data":"010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.251834 4788 generic.go:334] "Generic (PLEG): container finished" podID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerID="f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1" exitCode=0 Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.251893 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b76v" event={"ID":"ff6f90f1-145e-41d4-9529-67d7929598a5","Type":"ContainerDied","Data":"f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.264351 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerID="cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87" exitCode=0 Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.264549 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hggwc" event={"ID":"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6","Type":"ContainerDied","Data":"cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.276473 4788 generic.go:334] "Generic (PLEG): container finished" podID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerID="3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37" exitCode=0 Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.276715 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7brq" event={"ID":"8ef8ef60-95ef-4e3d-8e13-149f602fc08d","Type":"ContainerDied","Data":"3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.286736 4788 generic.go:334] "Generic (PLEG): container finished" podID="8308414f-3baf-4667-9b05-801f0fbde688" containerID="630fabae9e15482601b2a4022402b70b5d467a2db9a9a565ddba984ba7674b73" exitCode=0 Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.286875 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qwf8" event={"ID":"8308414f-3baf-4667-9b05-801f0fbde688","Type":"ContainerDied","Data":"630fabae9e15482601b2a4022402b70b5d467a2db9a9a565ddba984ba7674b73"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.295719 4788 generic.go:334] "Generic (PLEG): container finished" podID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerID="5fe194d114f06da7c83ea5e21969f4eec3793d43a44602612ea8643a5115ab3b" exitCode=0 Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.295800 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkwgg" event={"ID":"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100","Type":"ContainerDied","Data":"5fe194d114f06da7c83ea5e21969f4eec3793d43a44602612ea8643a5115ab3b"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.300272 4788 generic.go:334] "Generic (PLEG): container finished" podID="10662620-8659-4eb6-b84d-420c73f200bb" containerID="96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a" exitCode=0 Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.301501 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkxhd" event={"ID":"10662620-8659-4eb6-b84d-420c73f200bb","Type":"ContainerDied","Data":"96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a"} Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.410334 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:47:59 crc kubenswrapper[4788]: I1010 14:47:59.410413 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:47:59 crc kubenswrapper[4788]: E1010 14:47:59.582799 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-conmon-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache]" Oct 10 14:48:00 crc kubenswrapper[4788]: I1010 14:48:00.317463 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2cft7" Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.327410 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzskc" event={"ID":"5bec33de-d109-4518-9a02-a9a65382a53b","Type":"ContainerStarted","Data":"18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2"} Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.330337 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b76v" event={"ID":"ff6f90f1-145e-41d4-9529-67d7929598a5","Type":"ContainerStarted","Data":"957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351"} Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.333697 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hggwc" event={"ID":"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6","Type":"ContainerStarted","Data":"b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e"} Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.335087 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7brq" event={"ID":"8ef8ef60-95ef-4e3d-8e13-149f602fc08d","Type":"ContainerStarted","Data":"841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e"} Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.337345 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qwf8" event={"ID":"8308414f-3baf-4667-9b05-801f0fbde688","Type":"ContainerStarted","Data":"025ef21b7f45615358209ffe2a3c1494bd4a9dbe23f5b8b606164b6806dcbafb"} Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.344008 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkwgg" event={"ID":"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100","Type":"ContainerStarted","Data":"824767b6041c1c6836f4a45fe6c5cdaf3292508af0046fed3a23a104f78f153e"} Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.346681 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkxhd" event={"ID":"10662620-8659-4eb6-b84d-420c73f200bb","Type":"ContainerStarted","Data":"b058bc0a0cc0af3434678a1e3630b23992106102f9f87e0debc2c23874da7df3"} Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.374737 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-662r8" event={"ID":"9479f6c6-ef3d-4a69-affe-2f773e32b48f","Type":"ContainerStarted","Data":"368fb1a937c26c14d40400719740f2694178d79a16f04e30eac340daa3748705"} Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.390487 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vzskc" podStartSLOduration=3.445901095 podStartE2EDuration="34.390460592s" podCreationTimestamp="2025-10-10 14:47:29 +0000 UTC" firstStartedPulling="2025-10-10 14:47:31.732885583 +0000 UTC m=+154.182601131" lastFinishedPulling="2025-10-10 14:48:02.67744504 +0000 UTC m=+185.127160628" observedRunningTime="2025-10-10 14:48:03.363547016 +0000 UTC m=+185.813262604" watchObservedRunningTime="2025-10-10 14:48:03.390460592 +0000 UTC m=+185.840176140" Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.391348 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5b76v" podStartSLOduration=3.1338941670000002 podStartE2EDuration="37.391340813s" podCreationTimestamp="2025-10-10 14:47:26 +0000 UTC" firstStartedPulling="2025-10-10 14:47:28.560361192 +0000 UTC m=+151.010076730" lastFinishedPulling="2025-10-10 14:48:02.817807818 +0000 UTC m=+185.267523376" observedRunningTime="2025-10-10 14:48:03.380746879 +0000 UTC m=+185.830462427" watchObservedRunningTime="2025-10-10 14:48:03.391340813 +0000 UTC m=+185.841056361" Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.410159 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mkxhd" podStartSLOduration=2.372905653 podStartE2EDuration="35.410123014s" podCreationTimestamp="2025-10-10 14:47:28 +0000 UTC" firstStartedPulling="2025-10-10 14:47:29.67653555 +0000 UTC m=+152.126251098" lastFinishedPulling="2025-10-10 14:48:02.713752871 +0000 UTC m=+185.163468459" observedRunningTime="2025-10-10 14:48:03.40997841 +0000 UTC m=+185.859693958" watchObservedRunningTime="2025-10-10 14:48:03.410123014 +0000 UTC m=+185.859838562" Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.433964 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-662r8" podStartSLOduration=2.327787711 podStartE2EDuration="34.433937325s" podCreationTimestamp="2025-10-10 14:47:29 +0000 UTC" firstStartedPulling="2025-10-10 14:47:30.758134869 +0000 UTC m=+153.207850417" lastFinishedPulling="2025-10-10 14:48:02.864284483 +0000 UTC m=+185.314000031" observedRunningTime="2025-10-10 14:48:03.430997895 +0000 UTC m=+185.880713443" watchObservedRunningTime="2025-10-10 14:48:03.433937325 +0000 UTC m=+185.883652883" Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.456877 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hggwc" podStartSLOduration=3.3384398060000002 podStartE2EDuration="37.456851655s" podCreationTimestamp="2025-10-10 14:47:26 +0000 UTC" firstStartedPulling="2025-10-10 14:47:28.590107956 +0000 UTC m=+151.039823504" lastFinishedPulling="2025-10-10 14:48:02.708519765 +0000 UTC m=+185.158235353" observedRunningTime="2025-10-10 14:48:03.453506625 +0000 UTC m=+185.903222173" watchObservedRunningTime="2025-10-10 14:48:03.456851655 +0000 UTC m=+185.906567203" Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.468569 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jkwgg" podStartSLOduration=3.045700419 podStartE2EDuration="37.468547996s" podCreationTimestamp="2025-10-10 14:47:26 +0000 UTC" firstStartedPulling="2025-10-10 14:47:28.60445948 +0000 UTC m=+151.054175028" lastFinishedPulling="2025-10-10 14:48:03.027307057 +0000 UTC m=+185.477022605" observedRunningTime="2025-10-10 14:48:03.466509897 +0000 UTC m=+185.916225445" watchObservedRunningTime="2025-10-10 14:48:03.468547996 +0000 UTC m=+185.918263544" Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.504181 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t7brq" podStartSLOduration=2.373231011 podStartE2EDuration="35.50415369s" podCreationTimestamp="2025-10-10 14:47:28 +0000 UTC" firstStartedPulling="2025-10-10 14:47:29.648255632 +0000 UTC m=+152.097971190" lastFinishedPulling="2025-10-10 14:48:02.779178321 +0000 UTC m=+185.228893869" observedRunningTime="2025-10-10 14:48:03.499356966 +0000 UTC m=+185.949072514" watchObservedRunningTime="2025-10-10 14:48:03.50415369 +0000 UTC m=+185.953869238" Oct 10 14:48:03 crc kubenswrapper[4788]: I1010 14:48:03.535031 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6qwf8" podStartSLOduration=3.216189852 podStartE2EDuration="37.535000771s" podCreationTimestamp="2025-10-10 14:47:26 +0000 UTC" firstStartedPulling="2025-10-10 14:47:28.599133393 +0000 UTC m=+151.048848951" lastFinishedPulling="2025-10-10 14:48:02.917944322 +0000 UTC m=+185.367659870" observedRunningTime="2025-10-10 14:48:03.528880214 +0000 UTC m=+185.978595762" watchObservedRunningTime="2025-10-10 14:48:03.535000771 +0000 UTC m=+185.984716319" Oct 10 14:48:06 crc kubenswrapper[4788]: I1010 14:48:06.358448 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 10 14:48:06 crc kubenswrapper[4788]: I1010 14:48:06.541167 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:48:06 crc kubenswrapper[4788]: I1010 14:48:06.541237 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:48:06 crc kubenswrapper[4788]: I1010 14:48:06.840851 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.109012 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.109099 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.156339 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.205324 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.205374 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.255469 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.276441 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.276476 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:48:07 crc kubenswrapper[4788]: I1010 14:48:07.325328 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:48:08 crc kubenswrapper[4788]: I1010 14:48:08.717616 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:48:08 crc kubenswrapper[4788]: I1010 14:48:08.718028 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:48:08 crc kubenswrapper[4788]: I1010 14:48:08.761282 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:48:09 crc kubenswrapper[4788]: I1010 14:48:09.033323 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:48:09 crc kubenswrapper[4788]: I1010 14:48:09.033984 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:48:09 crc kubenswrapper[4788]: I1010 14:48:09.093533 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:48:09 crc kubenswrapper[4788]: I1010 14:48:09.443528 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:48:09 crc kubenswrapper[4788]: I1010 14:48:09.459419 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:48:09 crc kubenswrapper[4788]: E1010 14:48:09.724329 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-conmon-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache]" Oct 10 14:48:09 crc kubenswrapper[4788]: I1010 14:48:09.882941 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:48:09 crc kubenswrapper[4788]: I1010 14:48:09.883017 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:48:09 crc kubenswrapper[4788]: I1010 14:48:09.936478 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:48:10 crc kubenswrapper[4788]: I1010 14:48:10.244010 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:48:10 crc kubenswrapper[4788]: I1010 14:48:10.244785 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:48:10 crc kubenswrapper[4788]: I1010 14:48:10.293950 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:48:10 crc kubenswrapper[4788]: I1010 14:48:10.458789 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:48:10 crc kubenswrapper[4788]: I1010 14:48:10.496409 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:48:12 crc kubenswrapper[4788]: I1010 14:48:12.537063 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7brq"] Oct 10 14:48:12 crc kubenswrapper[4788]: I1010 14:48:12.537653 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t7brq" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerName="registry-server" containerID="cri-o://841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e" gracePeriod=2 Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.028090 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.099978 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-catalog-content\") pod \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.100063 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-utilities\") pod \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.100108 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxzc4\" (UniqueName: \"kubernetes.io/projected/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-kube-api-access-hxzc4\") pod \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\" (UID: \"8ef8ef60-95ef-4e3d-8e13-149f602fc08d\") " Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.102037 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-utilities" (OuterVolumeSpecName: "utilities") pod "8ef8ef60-95ef-4e3d-8e13-149f602fc08d" (UID: "8ef8ef60-95ef-4e3d-8e13-149f602fc08d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.106730 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-kube-api-access-hxzc4" (OuterVolumeSpecName: "kube-api-access-hxzc4") pod "8ef8ef60-95ef-4e3d-8e13-149f602fc08d" (UID: "8ef8ef60-95ef-4e3d-8e13-149f602fc08d"). InnerVolumeSpecName "kube-api-access-hxzc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.118535 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ef8ef60-95ef-4e3d-8e13-149f602fc08d" (UID: "8ef8ef60-95ef-4e3d-8e13-149f602fc08d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.138752 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vzskc"] Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.138996 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vzskc" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" containerName="registry-server" containerID="cri-o://18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2" gracePeriod=2 Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.201262 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.201314 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.201326 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxzc4\" (UniqueName: \"kubernetes.io/projected/8ef8ef60-95ef-4e3d-8e13-149f602fc08d-kube-api-access-hxzc4\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.433640 4788 generic.go:334] "Generic (PLEG): container finished" podID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerID="841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e" exitCode=0 Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.433712 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7brq" event={"ID":"8ef8ef60-95ef-4e3d-8e13-149f602fc08d","Type":"ContainerDied","Data":"841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e"} Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.433746 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7brq" event={"ID":"8ef8ef60-95ef-4e3d-8e13-149f602fc08d","Type":"ContainerDied","Data":"43c95117a18a896c5e7087075f3c71c3955b25fff50d850e6a93e2bc8e726e17"} Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.433773 4788 scope.go:117] "RemoveContainer" containerID="841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.433966 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7brq" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.519805 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7brq"] Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.524188 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7brq"] Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.524342 4788 scope.go:117] "RemoveContainer" containerID="3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.547900 4788 scope.go:117] "RemoveContainer" containerID="2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.569299 4788 scope.go:117] "RemoveContainer" containerID="841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e" Oct 10 14:48:13 crc kubenswrapper[4788]: E1010 14:48:13.571653 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e\": container with ID starting with 841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e not found: ID does not exist" containerID="841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.571705 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e"} err="failed to get container status \"841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e\": rpc error: code = NotFound desc = could not find container \"841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e\": container with ID starting with 841fd85d87b7edf609d8a0d8687c18a27abeac40540ae79e07e994db68b8156e not found: ID does not exist" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.571763 4788 scope.go:117] "RemoveContainer" containerID="3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37" Oct 10 14:48:13 crc kubenswrapper[4788]: E1010 14:48:13.572303 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37\": container with ID starting with 3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37 not found: ID does not exist" containerID="3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.572361 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37"} err="failed to get container status \"3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37\": rpc error: code = NotFound desc = could not find container \"3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37\": container with ID starting with 3a23d7e671fe47d436c3b0a3f605a4d47a4847f0c18fb7d51e11d07140c93c37 not found: ID does not exist" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.572400 4788 scope.go:117] "RemoveContainer" containerID="2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36" Oct 10 14:48:13 crc kubenswrapper[4788]: E1010 14:48:13.572800 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36\": container with ID starting with 2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36 not found: ID does not exist" containerID="2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36" Oct 10 14:48:13 crc kubenswrapper[4788]: I1010 14:48:13.572868 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36"} err="failed to get container status \"2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36\": rpc error: code = NotFound desc = could not find container \"2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36\": container with ID starting with 2508ad6cbd2582b1a9d33c8b22e96282c4582d6df9253d521b7b28899d690e36 not found: ID does not exist" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.241672 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" path="/var/lib/kubelet/pods/8ef8ef60-95ef-4e3d-8e13-149f602fc08d/volumes" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.341343 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.416924 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-utilities\") pod \"5bec33de-d109-4518-9a02-a9a65382a53b\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.416973 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-catalog-content\") pod \"5bec33de-d109-4518-9a02-a9a65382a53b\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.417042 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wn2f\" (UniqueName: \"kubernetes.io/projected/5bec33de-d109-4518-9a02-a9a65382a53b-kube-api-access-9wn2f\") pod \"5bec33de-d109-4518-9a02-a9a65382a53b\" (UID: \"5bec33de-d109-4518-9a02-a9a65382a53b\") " Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.417850 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-utilities" (OuterVolumeSpecName: "utilities") pod "5bec33de-d109-4518-9a02-a9a65382a53b" (UID: "5bec33de-d109-4518-9a02-a9a65382a53b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.426811 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bec33de-d109-4518-9a02-a9a65382a53b-kube-api-access-9wn2f" (OuterVolumeSpecName: "kube-api-access-9wn2f") pod "5bec33de-d109-4518-9a02-a9a65382a53b" (UID: "5bec33de-d109-4518-9a02-a9a65382a53b"). InnerVolumeSpecName "kube-api-access-9wn2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.453893 4788 generic.go:334] "Generic (PLEG): container finished" podID="5bec33de-d109-4518-9a02-a9a65382a53b" containerID="18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2" exitCode=0 Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.453971 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vzskc" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.453980 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzskc" event={"ID":"5bec33de-d109-4518-9a02-a9a65382a53b","Type":"ContainerDied","Data":"18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2"} Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.454011 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vzskc" event={"ID":"5bec33de-d109-4518-9a02-a9a65382a53b","Type":"ContainerDied","Data":"55a86c4757aa6848f91690aa4007a69397a1961e1621a5d09221fff39937fecc"} Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.454053 4788 scope.go:117] "RemoveContainer" containerID="18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.474472 4788 scope.go:117] "RemoveContainer" containerID="010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.497257 4788 scope.go:117] "RemoveContainer" containerID="64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.509104 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bec33de-d109-4518-9a02-a9a65382a53b" (UID: "5bec33de-d109-4518-9a02-a9a65382a53b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.518137 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.518199 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bec33de-d109-4518-9a02-a9a65382a53b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.518210 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wn2f\" (UniqueName: \"kubernetes.io/projected/5bec33de-d109-4518-9a02-a9a65382a53b-kube-api-access-9wn2f\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.518685 4788 scope.go:117] "RemoveContainer" containerID="18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2" Oct 10 14:48:14 crc kubenswrapper[4788]: E1010 14:48:14.519181 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2\": container with ID starting with 18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2 not found: ID does not exist" containerID="18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.519209 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2"} err="failed to get container status \"18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2\": rpc error: code = NotFound desc = could not find container \"18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2\": container with ID starting with 18ddbc15c15f56e2898ef7ee1540e18bffa1305d635bcde6a0b2edaa3b7ee2f2 not found: ID does not exist" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.519284 4788 scope.go:117] "RemoveContainer" containerID="010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746" Oct 10 14:48:14 crc kubenswrapper[4788]: E1010 14:48:14.519792 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746\": container with ID starting with 010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746 not found: ID does not exist" containerID="010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.519855 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746"} err="failed to get container status \"010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746\": rpc error: code = NotFound desc = could not find container \"010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746\": container with ID starting with 010a8e337f51e3f56f02966c36382c806a43f64700ebd12a5fd742e52befb746 not found: ID does not exist" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.519897 4788 scope.go:117] "RemoveContainer" containerID="64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e" Oct 10 14:48:14 crc kubenswrapper[4788]: E1010 14:48:14.520307 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e\": container with ID starting with 64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e not found: ID does not exist" containerID="64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.520355 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e"} err="failed to get container status \"64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e\": rpc error: code = NotFound desc = could not find container \"64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e\": container with ID starting with 64ba1f78bd7c9d16a68e450cf842629ce82f298c8c5ba33c549c97e391e6f93e not found: ID does not exist" Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.788250 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vzskc"] Oct 10 14:48:14 crc kubenswrapper[4788]: I1010 14:48:14.795628 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vzskc"] Oct 10 14:48:16 crc kubenswrapper[4788]: I1010 14:48:16.242447 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" path="/var/lib/kubelet/pods/5bec33de-d109-4518-9a02-a9a65382a53b/volumes" Oct 10 14:48:16 crc kubenswrapper[4788]: I1010 14:48:16.589544 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:48:17 crc kubenswrapper[4788]: I1010 14:48:17.171085 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:48:17 crc kubenswrapper[4788]: I1010 14:48:17.256547 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:48:17 crc kubenswrapper[4788]: I1010 14:48:17.319243 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:48:19 crc kubenswrapper[4788]: I1010 14:48:19.536967 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5b76v"] Oct 10 14:48:19 crc kubenswrapper[4788]: I1010 14:48:19.537821 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5b76v" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerName="registry-server" containerID="cri-o://957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351" gracePeriod=2 Oct 10 14:48:19 crc kubenswrapper[4788]: I1010 14:48:19.624544 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5j9sp"] Oct 10 14:48:19 crc kubenswrapper[4788]: E1010 14:48:19.900089 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-conmon-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache]" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.036852 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.107732 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-catalog-content\") pod \"ff6f90f1-145e-41d4-9529-67d7929598a5\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.108160 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-utilities\") pod \"ff6f90f1-145e-41d4-9529-67d7929598a5\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.108193 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4zz8\" (UniqueName: \"kubernetes.io/projected/ff6f90f1-145e-41d4-9529-67d7929598a5-kube-api-access-d4zz8\") pod \"ff6f90f1-145e-41d4-9529-67d7929598a5\" (UID: \"ff6f90f1-145e-41d4-9529-67d7929598a5\") " Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.108841 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-utilities" (OuterVolumeSpecName: "utilities") pod "ff6f90f1-145e-41d4-9529-67d7929598a5" (UID: "ff6f90f1-145e-41d4-9529-67d7929598a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.116053 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff6f90f1-145e-41d4-9529-67d7929598a5-kube-api-access-d4zz8" (OuterVolumeSpecName: "kube-api-access-d4zz8") pod "ff6f90f1-145e-41d4-9529-67d7929598a5" (UID: "ff6f90f1-145e-41d4-9529-67d7929598a5"). InnerVolumeSpecName "kube-api-access-d4zz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.152126 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff6f90f1-145e-41d4-9529-67d7929598a5" (UID: "ff6f90f1-145e-41d4-9529-67d7929598a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.210132 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4zz8\" (UniqueName: \"kubernetes.io/projected/ff6f90f1-145e-41d4-9529-67d7929598a5-kube-api-access-d4zz8\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.210524 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.210545 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff6f90f1-145e-41d4-9529-67d7929598a5-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.504625 4788 generic.go:334] "Generic (PLEG): container finished" podID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerID="957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351" exitCode=0 Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.504674 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b76v" event={"ID":"ff6f90f1-145e-41d4-9529-67d7929598a5","Type":"ContainerDied","Data":"957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351"} Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.504684 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5b76v" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.504707 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5b76v" event={"ID":"ff6f90f1-145e-41d4-9529-67d7929598a5","Type":"ContainerDied","Data":"1f4b4572e8c0898a357dc4b0a33444244aeb44b2011b1898a59efc320127b788"} Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.504724 4788 scope.go:117] "RemoveContainer" containerID="957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.524894 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5b76v"] Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.528742 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5b76v"] Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.547208 4788 scope.go:117] "RemoveContainer" containerID="f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.567931 4788 scope.go:117] "RemoveContainer" containerID="157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.582206 4788 scope.go:117] "RemoveContainer" containerID="957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351" Oct 10 14:48:20 crc kubenswrapper[4788]: E1010 14:48:20.582644 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351\": container with ID starting with 957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351 not found: ID does not exist" containerID="957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.582697 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351"} err="failed to get container status \"957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351\": rpc error: code = NotFound desc = could not find container \"957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351\": container with ID starting with 957e43f243df306b7f69ace6b6c5ad212875cafb60c5c9090fdb627ff0d1c351 not found: ID does not exist" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.582733 4788 scope.go:117] "RemoveContainer" containerID="f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1" Oct 10 14:48:20 crc kubenswrapper[4788]: E1010 14:48:20.583080 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1\": container with ID starting with f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1 not found: ID does not exist" containerID="f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.583120 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1"} err="failed to get container status \"f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1\": rpc error: code = NotFound desc = could not find container \"f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1\": container with ID starting with f48e2965e9b9896bbb0910698130d6c51c167932c01305644596366ed54fc7b1 not found: ID does not exist" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.583165 4788 scope.go:117] "RemoveContainer" containerID="157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8" Oct 10 14:48:20 crc kubenswrapper[4788]: E1010 14:48:20.583448 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8\": container with ID starting with 157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8 not found: ID does not exist" containerID="157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.583479 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8"} err="failed to get container status \"157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8\": rpc error: code = NotFound desc = could not find container \"157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8\": container with ID starting with 157eaf6c9b13ed51643df7b6ed1c8786d89636f16c2d25561c6ea44317def6d8 not found: ID does not exist" Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.938561 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hggwc"] Oct 10 14:48:20 crc kubenswrapper[4788]: I1010 14:48:20.939049 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hggwc" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerName="registry-server" containerID="cri-o://b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e" gracePeriod=2 Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.310740 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.436485 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-utilities\") pod \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.436566 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk2gv\" (UniqueName: \"kubernetes.io/projected/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-kube-api-access-dk2gv\") pod \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.436603 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-catalog-content\") pod \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\" (UID: \"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6\") " Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.438332 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-utilities" (OuterVolumeSpecName: "utilities") pod "e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" (UID: "e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.460792 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-kube-api-access-dk2gv" (OuterVolumeSpecName: "kube-api-access-dk2gv") pod "e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" (UID: "e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6"). InnerVolumeSpecName "kube-api-access-dk2gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.513779 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerID="b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e" exitCode=0 Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.513844 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hggwc" event={"ID":"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6","Type":"ContainerDied","Data":"b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e"} Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.513883 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hggwc" event={"ID":"e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6","Type":"ContainerDied","Data":"ccc28281296d8f8567b8b72a8aa273fe8b3499e6a5404fe8006ea0806b8a3a56"} Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.513904 4788 scope.go:117] "RemoveContainer" containerID="b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.513929 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hggwc" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.527549 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" (UID: "e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.529270 4788 scope.go:117] "RemoveContainer" containerID="cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.538262 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.538296 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.538310 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk2gv\" (UniqueName: \"kubernetes.io/projected/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6-kube-api-access-dk2gv\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.547503 4788 scope.go:117] "RemoveContainer" containerID="df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.561566 4788 scope.go:117] "RemoveContainer" containerID="b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e" Oct 10 14:48:21 crc kubenswrapper[4788]: E1010 14:48:21.562599 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e\": container with ID starting with b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e not found: ID does not exist" containerID="b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.562644 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e"} err="failed to get container status \"b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e\": rpc error: code = NotFound desc = could not find container \"b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e\": container with ID starting with b9598cbe0f39d2c08e721f709c7c15ee569ffbb4f0db6f2af533847ea99faa9e not found: ID does not exist" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.562674 4788 scope.go:117] "RemoveContainer" containerID="cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87" Oct 10 14:48:21 crc kubenswrapper[4788]: E1010 14:48:21.562987 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87\": container with ID starting with cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87 not found: ID does not exist" containerID="cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.563038 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87"} err="failed to get container status \"cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87\": rpc error: code = NotFound desc = could not find container \"cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87\": container with ID starting with cd1bdea5893099222479fad07a7cf2f2f29ea51aac5b4f3ffcc845fff72d9c87 not found: ID does not exist" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.563068 4788 scope.go:117] "RemoveContainer" containerID="df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579" Oct 10 14:48:21 crc kubenswrapper[4788]: E1010 14:48:21.563571 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579\": container with ID starting with df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579 not found: ID does not exist" containerID="df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.563646 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579"} err="failed to get container status \"df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579\": rpc error: code = NotFound desc = could not find container \"df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579\": container with ID starting with df50fee900d585c2256a5fb4cd6cc7a5c8d70751df5d0d2a0300a5185135c579 not found: ID does not exist" Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.847371 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hggwc"] Oct 10 14:48:21 crc kubenswrapper[4788]: I1010 14:48:21.851466 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hggwc"] Oct 10 14:48:22 crc kubenswrapper[4788]: I1010 14:48:22.243068 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" path="/var/lib/kubelet/pods/e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6/volumes" Oct 10 14:48:22 crc kubenswrapper[4788]: I1010 14:48:22.243689 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" path="/var/lib/kubelet/pods/ff6f90f1-145e-41d4-9529-67d7929598a5/volumes" Oct 10 14:48:29 crc kubenswrapper[4788]: I1010 14:48:29.406229 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:48:29 crc kubenswrapper[4788]: I1010 14:48:29.406914 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:48:29 crc kubenswrapper[4788]: I1010 14:48:29.406967 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:48:29 crc kubenswrapper[4788]: I1010 14:48:29.407503 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:48:29 crc kubenswrapper[4788]: I1010 14:48:29.407581 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817" gracePeriod=600 Oct 10 14:48:29 crc kubenswrapper[4788]: I1010 14:48:29.587968 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817" exitCode=0 Oct 10 14:48:29 crc kubenswrapper[4788]: I1010 14:48:29.588058 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817"} Oct 10 14:48:30 crc kubenswrapper[4788]: E1010 14:48:30.044054 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-conmon-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache]" Oct 10 14:48:30 crc kubenswrapper[4788]: I1010 14:48:30.597480 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"cc73d4024eb0529c2e8485e4f790187ca0d08a7d42f2bdd40bacb41d592bece3"} Oct 10 14:48:40 crc kubenswrapper[4788]: E1010 14:48:40.210376 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-conmon-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache]" Oct 10 14:48:44 crc kubenswrapper[4788]: I1010 14:48:44.654134 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" podUID="3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" containerName="oauth-openshift" containerID="cri-o://ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e" gracePeriod=15 Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.077926 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115427 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d6876856f-7h49f"] Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115641 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115652 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115666 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerName="extract-utilities" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115672 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerName="extract-utilities" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115683 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" containerName="oauth-openshift" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115690 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" containerName="oauth-openshift" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115697 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerName="extract-content" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115705 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerName="extract-content" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115712 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115718 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115728 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115735 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115744 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a630608-9824-4339-a050-8c1c83b88b5b" containerName="pruner" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115750 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a630608-9824-4339-a050-8c1c83b88b5b" containerName="pruner" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115762 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8612809d-f9ea-4c98-9eea-26f9540d8613" containerName="pruner" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115768 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8612809d-f9ea-4c98-9eea-26f9540d8613" containerName="pruner" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115775 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" containerName="extract-content" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115780 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" containerName="extract-content" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115788 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerName="extract-utilities" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115794 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerName="extract-utilities" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115801 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerName="extract-utilities" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115807 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerName="extract-utilities" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115815 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" containerName="extract-utilities" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115821 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" containerName="extract-utilities" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115830 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerName="extract-content" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115836 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerName="extract-content" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115843 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115848 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.115855 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerName="extract-content" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115862 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerName="extract-content" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115941 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a630608-9824-4339-a050-8c1c83b88b5b" containerName="pruner" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115953 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ef8ef60-95ef-4e3d-8e13-149f602fc08d" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115962 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8612809d-f9ea-4c98-9eea-26f9540d8613" containerName="pruner" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115971 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9b85b1a-ac5c-4aad-8183-dcd0fb4cf5c6" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115979 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bec33de-d109-4518-9a02-a9a65382a53b" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115986 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff6f90f1-145e-41d4-9529-67d7929598a5" containerName="registry-server" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.115996 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" containerName="oauth-openshift" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.116426 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.130536 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d6876856f-7h49f"] Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.170933 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-login\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171007 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-trusted-ca-bundle\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171049 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-cliconfig\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171081 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-error\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171118 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-service-ca\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171174 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-policies\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171203 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-idp-0-file-data\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171240 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-router-certs\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171292 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-session\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171323 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9pz2\" (UniqueName: \"kubernetes.io/projected/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-kube-api-access-w9pz2\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171350 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-dir\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171418 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-ocp-branding-template\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171487 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-serving-cert\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171513 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-provider-selection\") pod \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\" (UID: \"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05\") " Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171787 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171853 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171881 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-login\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171902 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-session\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171953 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.171988 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172021 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172045 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172206 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172285 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-error\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172395 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txnbv\" (UniqueName: \"kubernetes.io/projected/7bb1ffd4-c1b7-439e-9b83-0999c14be224-kube-api-access-txnbv\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172429 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172473 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172497 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-audit-policies\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172562 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bb1ffd4-c1b7-439e-9b83-0999c14be224-audit-dir\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172666 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172735 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.172749 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.173545 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.173887 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.174248 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.178244 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.178582 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.186980 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-kube-api-access-w9pz2" (OuterVolumeSpecName: "kube-api-access-w9pz2") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "kube-api-access-w9pz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.188467 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.188741 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.188940 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.189360 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.189386 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.189572 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" (UID: "3a56998c-1bb2-4c0e-b5b0-d8a1be704e05"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.274195 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.274500 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.274607 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-audit-policies\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.274777 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bb1ffd4-c1b7-439e-9b83-0999c14be224-audit-dir\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.274886 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.274931 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bb1ffd4-c1b7-439e-9b83-0999c14be224-audit-dir\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.274983 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275133 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275183 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-login\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275208 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-session\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275233 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275307 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275332 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275387 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-error\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275462 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txnbv\" (UniqueName: \"kubernetes.io/projected/7bb1ffd4-c1b7-439e-9b83-0999c14be224-kube-api-access-txnbv\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275746 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275768 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275779 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275790 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275802 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275895 4788 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275906 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275917 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275928 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275937 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9pz2\" (UniqueName: \"kubernetes.io/projected/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-kube-api-access-w9pz2\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275946 4788 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.275956 4788 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.276094 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.276250 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.276369 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-audit-policies\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.277267 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.279185 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-error\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.279210 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-login\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.280425 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.280804 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.281131 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-session\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.281530 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.281723 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.283809 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bb1ffd4-c1b7-439e-9b83-0999c14be224-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.303450 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txnbv\" (UniqueName: \"kubernetes.io/projected/7bb1ffd4-c1b7-439e-9b83-0999c14be224-kube-api-access-txnbv\") pod \"oauth-openshift-7d6876856f-7h49f\" (UID: \"7bb1ffd4-c1b7-439e-9b83-0999c14be224\") " pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.438805 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.687444 4788 generic.go:334] "Generic (PLEG): container finished" podID="3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" containerID="ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e" exitCode=0 Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.687530 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" event={"ID":"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05","Type":"ContainerDied","Data":"ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e"} Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.687591 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" event={"ID":"3a56998c-1bb2-4c0e-b5b0-d8a1be704e05","Type":"ContainerDied","Data":"f3fed2cc75e0676d281b6a80c6c408004d67492a4d370ab1492de4079c03ee26"} Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.687591 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5j9sp" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.687627 4788 scope.go:117] "RemoveContainer" containerID="ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.706015 4788 scope.go:117] "RemoveContainer" containerID="ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e" Oct 10 14:48:45 crc kubenswrapper[4788]: E1010 14:48:45.706541 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e\": container with ID starting with ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e not found: ID does not exist" containerID="ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.706597 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e"} err="failed to get container status \"ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e\": rpc error: code = NotFound desc = could not find container \"ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e\": container with ID starting with ad2add4a62f32c4680480817b44f19c6510ae83b8b4f9a642b8421c5c295cb3e not found: ID does not exist" Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.731670 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5j9sp"] Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.733269 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5j9sp"] Oct 10 14:48:45 crc kubenswrapper[4788]: I1010 14:48:45.886184 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d6876856f-7h49f"] Oct 10 14:48:45 crc kubenswrapper[4788]: W1010 14:48:45.892808 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bb1ffd4_c1b7_439e_9b83_0999c14be224.slice/crio-ab9727520aca5d242072ee31b48bbc31a2766055d374f197ae0b43f024687467 WatchSource:0}: Error finding container ab9727520aca5d242072ee31b48bbc31a2766055d374f197ae0b43f024687467: Status 404 returned error can't find the container with id ab9727520aca5d242072ee31b48bbc31a2766055d374f197ae0b43f024687467 Oct 10 14:48:46 crc kubenswrapper[4788]: I1010 14:48:46.246719 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a56998c-1bb2-4c0e-b5b0-d8a1be704e05" path="/var/lib/kubelet/pods/3a56998c-1bb2-4c0e-b5b0-d8a1be704e05/volumes" Oct 10 14:48:46 crc kubenswrapper[4788]: I1010 14:48:46.697994 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" event={"ID":"7bb1ffd4-c1b7-439e-9b83-0999c14be224","Type":"ContainerStarted","Data":"296d511f8e6ee5448f430d636cf5b761269005804c5051c90b42b5bf6f766f8b"} Oct 10 14:48:46 crc kubenswrapper[4788]: I1010 14:48:46.698041 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" event={"ID":"7bb1ffd4-c1b7-439e-9b83-0999c14be224","Type":"ContainerStarted","Data":"ab9727520aca5d242072ee31b48bbc31a2766055d374f197ae0b43f024687467"} Oct 10 14:48:46 crc kubenswrapper[4788]: I1010 14:48:46.698520 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:46 crc kubenswrapper[4788]: I1010 14:48:46.706942 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" Oct 10 14:48:46 crc kubenswrapper[4788]: I1010 14:48:46.721993 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d6876856f-7h49f" podStartSLOduration=27.721972191 podStartE2EDuration="27.721972191s" podCreationTimestamp="2025-10-10 14:48:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:48:46.721281075 +0000 UTC m=+229.170996633" watchObservedRunningTime="2025-10-10 14:48:46.721972191 +0000 UTC m=+229.171687739" Oct 10 14:48:50 crc kubenswrapper[4788]: E1010 14:48:50.344305 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-conmon-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10662620_8659_4eb6_b84d_420c73f200bb.slice/crio-96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a.scope\": RecentStats: unable to find data in memory cache]" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.328381 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jkwgg"] Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.329539 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jkwgg" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerName="registry-server" containerID="cri-o://824767b6041c1c6836f4a45fe6c5cdaf3292508af0046fed3a23a104f78f153e" gracePeriod=30 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.350266 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6qwf8"] Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.350743 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6qwf8" podUID="8308414f-3baf-4667-9b05-801f0fbde688" containerName="registry-server" containerID="cri-o://025ef21b7f45615358209ffe2a3c1494bd4a9dbe23f5b8b606164b6806dcbafb" gracePeriod=30 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.358113 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nrzg2"] Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.359654 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" podUID="2686886b-00d1-4e24-8a84-f634c766fcca" containerName="marketplace-operator" containerID="cri-o://bd354c6be945783100ba01d4b72112ac9d919792a526513c514770caf3210e61" gracePeriod=30 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.365783 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkxhd"] Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.366079 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mkxhd" podUID="10662620-8659-4eb6-b84d-420c73f200bb" containerName="registry-server" containerID="cri-o://b058bc0a0cc0af3434678a1e3630b23992106102f9f87e0debc2c23874da7df3" gracePeriod=30 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.372102 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pd5nn"] Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.372796 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.394594 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pd5nn"] Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.397535 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-662r8"] Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.397811 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-662r8" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerName="registry-server" containerID="cri-o://368fb1a937c26c14d40400719740f2694178d79a16f04e30eac340daa3748705" gracePeriod=30 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.430152 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/04d1b9a1-3e55-480c-aab6-257b2095fd7d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.430307 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04d1b9a1-3e55-480c-aab6-257b2095fd7d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.430412 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8lw7\" (UniqueName: \"kubernetes.io/projected/04d1b9a1-3e55-480c-aab6-257b2095fd7d-kube-api-access-r8lw7\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.531630 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/04d1b9a1-3e55-480c-aab6-257b2095fd7d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.531706 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04d1b9a1-3e55-480c-aab6-257b2095fd7d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.531737 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8lw7\" (UniqueName: \"kubernetes.io/projected/04d1b9a1-3e55-480c-aab6-257b2095fd7d-kube-api-access-r8lw7\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.534319 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04d1b9a1-3e55-480c-aab6-257b2095fd7d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.540123 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/04d1b9a1-3e55-480c-aab6-257b2095fd7d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.549013 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8lw7\" (UniqueName: \"kubernetes.io/projected/04d1b9a1-3e55-480c-aab6-257b2095fd7d-kube-api-access-r8lw7\") pod \"marketplace-operator-79b997595-pd5nn\" (UID: \"04d1b9a1-3e55-480c-aab6-257b2095fd7d\") " pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.689082 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.762634 4788 generic.go:334] "Generic (PLEG): container finished" podID="8308414f-3baf-4667-9b05-801f0fbde688" containerID="025ef21b7f45615358209ffe2a3c1494bd4a9dbe23f5b8b606164b6806dcbafb" exitCode=0 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.762706 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qwf8" event={"ID":"8308414f-3baf-4667-9b05-801f0fbde688","Type":"ContainerDied","Data":"025ef21b7f45615358209ffe2a3c1494bd4a9dbe23f5b8b606164b6806dcbafb"} Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.774458 4788 generic.go:334] "Generic (PLEG): container finished" podID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerID="824767b6041c1c6836f4a45fe6c5cdaf3292508af0046fed3a23a104f78f153e" exitCode=0 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.774515 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkwgg" event={"ID":"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100","Type":"ContainerDied","Data":"824767b6041c1c6836f4a45fe6c5cdaf3292508af0046fed3a23a104f78f153e"} Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.774540 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkwgg" event={"ID":"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100","Type":"ContainerDied","Data":"0fc4c9f421e9214a63b17095e0597c011e0b7967e18cc98a9170465c68f1d8bf"} Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.774553 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fc4c9f421e9214a63b17095e0597c011e0b7967e18cc98a9170465c68f1d8bf" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.776667 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.777572 4788 generic.go:334] "Generic (PLEG): container finished" podID="10662620-8659-4eb6-b84d-420c73f200bb" containerID="b058bc0a0cc0af3434678a1e3630b23992106102f9f87e0debc2c23874da7df3" exitCode=0 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.777620 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkxhd" event={"ID":"10662620-8659-4eb6-b84d-420c73f200bb","Type":"ContainerDied","Data":"b058bc0a0cc0af3434678a1e3630b23992106102f9f87e0debc2c23874da7df3"} Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.807434 4788 generic.go:334] "Generic (PLEG): container finished" podID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerID="368fb1a937c26c14d40400719740f2694178d79a16f04e30eac340daa3748705" exitCode=0 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.807534 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-662r8" event={"ID":"9479f6c6-ef3d-4a69-affe-2f773e32b48f","Type":"ContainerDied","Data":"368fb1a937c26c14d40400719740f2694178d79a16f04e30eac340daa3748705"} Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.827757 4788 generic.go:334] "Generic (PLEG): container finished" podID="2686886b-00d1-4e24-8a84-f634c766fcca" containerID="bd354c6be945783100ba01d4b72112ac9d919792a526513c514770caf3210e61" exitCode=0 Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.827810 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" event={"ID":"2686886b-00d1-4e24-8a84-f634c766fcca","Type":"ContainerDied","Data":"bd354c6be945783100ba01d4b72112ac9d919792a526513c514770caf3210e61"} Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.835530 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fd57\" (UniqueName: \"kubernetes.io/projected/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-kube-api-access-7fd57\") pod \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.835694 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-utilities\") pod \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.835755 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-catalog-content\") pod \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\" (UID: \"5fd4a6f4-51f2-4d9a-88d2-a68c7270a100\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.841712 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-utilities" (OuterVolumeSpecName: "utilities") pod "5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" (UID: "5fd4a6f4-51f2-4d9a-88d2-a68c7270a100"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.842766 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-kube-api-access-7fd57" (OuterVolumeSpecName: "kube-api-access-7fd57") pod "5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" (UID: "5fd4a6f4-51f2-4d9a-88d2-a68c7270a100"). InnerVolumeSpecName "kube-api-access-7fd57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.858005 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.881426 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.901710 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.906439 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" (UID: "5fd4a6f4-51f2-4d9a-88d2-a68c7270a100"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.912092 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936581 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq4dz\" (UniqueName: \"kubernetes.io/projected/8308414f-3baf-4667-9b05-801f0fbde688-kube-api-access-fq4dz\") pod \"8308414f-3baf-4667-9b05-801f0fbde688\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936630 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-catalog-content\") pod \"8308414f-3baf-4667-9b05-801f0fbde688\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936682 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv4nr\" (UniqueName: \"kubernetes.io/projected/10662620-8659-4eb6-b84d-420c73f200bb-kube-api-access-dv4nr\") pod \"10662620-8659-4eb6-b84d-420c73f200bb\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936705 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-utilities\") pod \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936735 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-catalog-content\") pod \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936758 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqkm5\" (UniqueName: \"kubernetes.io/projected/9479f6c6-ef3d-4a69-affe-2f773e32b48f-kube-api-access-jqkm5\") pod \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\" (UID: \"9479f6c6-ef3d-4a69-affe-2f773e32b48f\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936783 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-operator-metrics\") pod \"2686886b-00d1-4e24-8a84-f634c766fcca\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936806 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-utilities\") pod \"10662620-8659-4eb6-b84d-420c73f200bb\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936837 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-trusted-ca\") pod \"2686886b-00d1-4e24-8a84-f634c766fcca\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936869 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlss6\" (UniqueName: \"kubernetes.io/projected/2686886b-00d1-4e24-8a84-f634c766fcca-kube-api-access-wlss6\") pod \"2686886b-00d1-4e24-8a84-f634c766fcca\" (UID: \"2686886b-00d1-4e24-8a84-f634c766fcca\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936890 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-utilities\") pod \"8308414f-3baf-4667-9b05-801f0fbde688\" (UID: \"8308414f-3baf-4667-9b05-801f0fbde688\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.936919 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-catalog-content\") pod \"10662620-8659-4eb6-b84d-420c73f200bb\" (UID: \"10662620-8659-4eb6-b84d-420c73f200bb\") " Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.937133 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.937165 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.937178 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fd57\" (UniqueName: \"kubernetes.io/projected/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100-kube-api-access-7fd57\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.939567 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2686886b-00d1-4e24-8a84-f634c766fcca" (UID: "2686886b-00d1-4e24-8a84-f634c766fcca"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.941110 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-utilities" (OuterVolumeSpecName: "utilities") pod "8308414f-3baf-4667-9b05-801f0fbde688" (UID: "8308414f-3baf-4667-9b05-801f0fbde688"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.941445 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8308414f-3baf-4667-9b05-801f0fbde688-kube-api-access-fq4dz" (OuterVolumeSpecName: "kube-api-access-fq4dz") pod "8308414f-3baf-4667-9b05-801f0fbde688" (UID: "8308414f-3baf-4667-9b05-801f0fbde688"). InnerVolumeSpecName "kube-api-access-fq4dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.941900 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-utilities" (OuterVolumeSpecName: "utilities") pod "9479f6c6-ef3d-4a69-affe-2f773e32b48f" (UID: "9479f6c6-ef3d-4a69-affe-2f773e32b48f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.942034 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-utilities" (OuterVolumeSpecName: "utilities") pod "10662620-8659-4eb6-b84d-420c73f200bb" (UID: "10662620-8659-4eb6-b84d-420c73f200bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.946324 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10662620-8659-4eb6-b84d-420c73f200bb-kube-api-access-dv4nr" (OuterVolumeSpecName: "kube-api-access-dv4nr") pod "10662620-8659-4eb6-b84d-420c73f200bb" (UID: "10662620-8659-4eb6-b84d-420c73f200bb"). InnerVolumeSpecName "kube-api-access-dv4nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.948038 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2686886b-00d1-4e24-8a84-f634c766fcca-kube-api-access-wlss6" (OuterVolumeSpecName: "kube-api-access-wlss6") pod "2686886b-00d1-4e24-8a84-f634c766fcca" (UID: "2686886b-00d1-4e24-8a84-f634c766fcca"). InnerVolumeSpecName "kube-api-access-wlss6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.948211 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9479f6c6-ef3d-4a69-affe-2f773e32b48f-kube-api-access-jqkm5" (OuterVolumeSpecName: "kube-api-access-jqkm5") pod "9479f6c6-ef3d-4a69-affe-2f773e32b48f" (UID: "9479f6c6-ef3d-4a69-affe-2f773e32b48f"). InnerVolumeSpecName "kube-api-access-jqkm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.957230 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2686886b-00d1-4e24-8a84-f634c766fcca" (UID: "2686886b-00d1-4e24-8a84-f634c766fcca"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:48:57 crc kubenswrapper[4788]: I1010 14:48:57.971672 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10662620-8659-4eb6-b84d-420c73f200bb" (UID: "10662620-8659-4eb6-b84d-420c73f200bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.017624 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8308414f-3baf-4667-9b05-801f0fbde688" (UID: "8308414f-3baf-4667-9b05-801f0fbde688"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.038526 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.038797 4788 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.038897 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlss6\" (UniqueName: \"kubernetes.io/projected/2686886b-00d1-4e24-8a84-f634c766fcca-kube-api-access-wlss6\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.038970 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.039027 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10662620-8659-4eb6-b84d-420c73f200bb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.039083 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq4dz\" (UniqueName: \"kubernetes.io/projected/8308414f-3baf-4667-9b05-801f0fbde688-kube-api-access-fq4dz\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.039137 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8308414f-3baf-4667-9b05-801f0fbde688-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.039215 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv4nr\" (UniqueName: \"kubernetes.io/projected/10662620-8659-4eb6-b84d-420c73f200bb-kube-api-access-dv4nr\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.039286 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.039349 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqkm5\" (UniqueName: \"kubernetes.io/projected/9479f6c6-ef3d-4a69-affe-2f773e32b48f-kube-api-access-jqkm5\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.039413 4788 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2686886b-00d1-4e24-8a84-f634c766fcca-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.046472 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9479f6c6-ef3d-4a69-affe-2f773e32b48f" (UID: "9479f6c6-ef3d-4a69-affe-2f773e32b48f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.141165 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9479f6c6-ef3d-4a69-affe-2f773e32b48f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.186498 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pd5nn"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.835654 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" event={"ID":"2686886b-00d1-4e24-8a84-f634c766fcca","Type":"ContainerDied","Data":"955e6795adf937975d3cb7afa4aa8c1bad75561298ea0540caa87244da9ee037"} Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.836438 4788 scope.go:117] "RemoveContainer" containerID="bd354c6be945783100ba01d4b72112ac9d919792a526513c514770caf3210e61" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.835869 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nrzg2" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.839676 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6qwf8" event={"ID":"8308414f-3baf-4667-9b05-801f0fbde688","Type":"ContainerDied","Data":"20ab22c578d6d1575a10d3f78e2b942546c037637ea8e18178969896e6523ab3"} Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.839790 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6qwf8" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.843563 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" event={"ID":"04d1b9a1-3e55-480c-aab6-257b2095fd7d","Type":"ContainerStarted","Data":"20d0de01889e7916cb1cc7f8db06c5ab35bef7932fc6eed1a36f966c1dcebeb1"} Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.843608 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" event={"ID":"04d1b9a1-3e55-480c-aab6-257b2095fd7d","Type":"ContainerStarted","Data":"6eaaeeac5e86363d5bbc62c83498818adb5c03bc7f659773cb6cadb6deba1688"} Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.844084 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.850021 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.852047 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkxhd" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.852677 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkxhd" event={"ID":"10662620-8659-4eb6-b84d-420c73f200bb","Type":"ContainerDied","Data":"26bbdb6a3a54d23985a644fe7b256b5cce67186558ceaf0952a5f611836129ec"} Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.857039 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nrzg2"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.859741 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nrzg2"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.861024 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkwgg" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.861340 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-662r8" event={"ID":"9479f6c6-ef3d-4a69-affe-2f773e32b48f","Type":"ContainerDied","Data":"1fd4aec502fc6e0a085d1e29e394ed137469eb183d2949dae8d0928b32e13dba"} Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.861565 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-662r8" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.866057 4788 scope.go:117] "RemoveContainer" containerID="025ef21b7f45615358209ffe2a3c1494bd4a9dbe23f5b8b606164b6806dcbafb" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.875055 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-pd5nn" podStartSLOduration=1.875025868 podStartE2EDuration="1.875025868s" podCreationTimestamp="2025-10-10 14:48:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:48:58.872806563 +0000 UTC m=+241.322522111" watchObservedRunningTime="2025-10-10 14:48:58.875025868 +0000 UTC m=+241.324741416" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.892891 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6qwf8"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.897345 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6qwf8"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.899277 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jkwgg"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.901625 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jkwgg"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.902278 4788 scope.go:117] "RemoveContainer" containerID="630fabae9e15482601b2a4022402b70b5d467a2db9a9a565ddba984ba7674b73" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.924546 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkxhd"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.926777 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkxhd"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.934273 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-662r8"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.939214 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-662r8"] Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.975470 4788 scope.go:117] "RemoveContainer" containerID="dc73e01b5493baef27fda9bfd6257d4b95bca0665cf39b2aa830a0ed6236385a" Oct 10 14:48:58 crc kubenswrapper[4788]: I1010 14:48:58.997917 4788 scope.go:117] "RemoveContainer" containerID="b058bc0a0cc0af3434678a1e3630b23992106102f9f87e0debc2c23874da7df3" Oct 10 14:48:59 crc kubenswrapper[4788]: I1010 14:48:59.022927 4788 scope.go:117] "RemoveContainer" containerID="96736a68948ba4a0daa8b15a48ff106ee5d965bb870e00ce47706535916faf3a" Oct 10 14:48:59 crc kubenswrapper[4788]: I1010 14:48:59.039593 4788 scope.go:117] "RemoveContainer" containerID="939fe39905084927723b11c348b2e4f8d197212ac85f9c86c776a179b4f8f48e" Oct 10 14:48:59 crc kubenswrapper[4788]: I1010 14:48:59.072697 4788 scope.go:117] "RemoveContainer" containerID="368fb1a937c26c14d40400719740f2694178d79a16f04e30eac340daa3748705" Oct 10 14:48:59 crc kubenswrapper[4788]: I1010 14:48:59.097242 4788 scope.go:117] "RemoveContainer" containerID="23a5a08ccea022a599f9335fe459f5892e9afef159120ca632f864dc9df311a1" Oct 10 14:48:59 crc kubenswrapper[4788]: I1010 14:48:59.113864 4788 scope.go:117] "RemoveContainer" containerID="2dec899db81718eb3d070f1d2a3cbed752f3853591457578060270a11237b541" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149546 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lz6z2"] Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149736 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149747 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149756 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2686886b-00d1-4e24-8a84-f634c766fcca" containerName="marketplace-operator" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149763 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2686886b-00d1-4e24-8a84-f634c766fcca" containerName="marketplace-operator" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149773 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerName="extract-utilities" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149781 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerName="extract-utilities" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149791 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerName="extract-content" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149798 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerName="extract-content" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149806 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerName="extract-content" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149812 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerName="extract-content" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149821 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8308414f-3baf-4667-9b05-801f0fbde688" containerName="extract-content" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149827 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8308414f-3baf-4667-9b05-801f0fbde688" containerName="extract-content" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149834 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10662620-8659-4eb6-b84d-420c73f200bb" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149840 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="10662620-8659-4eb6-b84d-420c73f200bb" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149847 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149854 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149863 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8308414f-3baf-4667-9b05-801f0fbde688" containerName="extract-utilities" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149869 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8308414f-3baf-4667-9b05-801f0fbde688" containerName="extract-utilities" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149878 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10662620-8659-4eb6-b84d-420c73f200bb" containerName="extract-content" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149884 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="10662620-8659-4eb6-b84d-420c73f200bb" containerName="extract-content" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149894 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10662620-8659-4eb6-b84d-420c73f200bb" containerName="extract-utilities" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149899 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="10662620-8659-4eb6-b84d-420c73f200bb" containerName="extract-utilities" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149907 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerName="extract-utilities" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149913 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerName="extract-utilities" Oct 10 14:49:00 crc kubenswrapper[4788]: E1010 14:49:00.149923 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8308414f-3baf-4667-9b05-801f0fbde688" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.149929 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8308414f-3baf-4667-9b05-801f0fbde688" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.150015 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="2686886b-00d1-4e24-8a84-f634c766fcca" containerName="marketplace-operator" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.150024 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="10662620-8659-4eb6-b84d-420c73f200bb" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.150032 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.150043 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.150052 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8308414f-3baf-4667-9b05-801f0fbde688" containerName="registry-server" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.150721 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.153376 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.167389 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lz6z2"] Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.197697 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-catalog-content\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.197773 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcsw7\" (UniqueName: \"kubernetes.io/projected/f4845d06-1130-4527-a881-c01fbec64e87-kube-api-access-bcsw7\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.197802 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-utilities\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.242172 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10662620-8659-4eb6-b84d-420c73f200bb" path="/var/lib/kubelet/pods/10662620-8659-4eb6-b84d-420c73f200bb/volumes" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.243546 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2686886b-00d1-4e24-8a84-f634c766fcca" path="/var/lib/kubelet/pods/2686886b-00d1-4e24-8a84-f634c766fcca/volumes" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.244708 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fd4a6f4-51f2-4d9a-88d2-a68c7270a100" path="/var/lib/kubelet/pods/5fd4a6f4-51f2-4d9a-88d2-a68c7270a100/volumes" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.246419 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8308414f-3baf-4667-9b05-801f0fbde688" path="/var/lib/kubelet/pods/8308414f-3baf-4667-9b05-801f0fbde688/volumes" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.247194 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9479f6c6-ef3d-4a69-affe-2f773e32b48f" path="/var/lib/kubelet/pods/9479f6c6-ef3d-4a69-affe-2f773e32b48f/volumes" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.298890 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-catalog-content\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.298956 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcsw7\" (UniqueName: \"kubernetes.io/projected/f4845d06-1130-4527-a881-c01fbec64e87-kube-api-access-bcsw7\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.298988 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-utilities\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.299572 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-utilities\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.299964 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-catalog-content\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.335340 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcsw7\" (UniqueName: \"kubernetes.io/projected/f4845d06-1130-4527-a881-c01fbec64e87-kube-api-access-bcsw7\") pod \"certified-operators-lz6z2\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.512345 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:00 crc kubenswrapper[4788]: I1010 14:49:00.959473 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lz6z2"] Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.149428 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q9598"] Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.150612 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.153838 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.159963 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9598"] Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.213864 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcf1e71b-3d94-4450-ae39-ea10deac5dde-catalog-content\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.214838 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlgvk\" (UniqueName: \"kubernetes.io/projected/dcf1e71b-3d94-4450-ae39-ea10deac5dde-kube-api-access-wlgvk\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.214873 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcf1e71b-3d94-4450-ae39-ea10deac5dde-utilities\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.315933 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlgvk\" (UniqueName: \"kubernetes.io/projected/dcf1e71b-3d94-4450-ae39-ea10deac5dde-kube-api-access-wlgvk\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.315997 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcf1e71b-3d94-4450-ae39-ea10deac5dde-utilities\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.316107 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcf1e71b-3d94-4450-ae39-ea10deac5dde-catalog-content\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.316905 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcf1e71b-3d94-4450-ae39-ea10deac5dde-catalog-content\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.317610 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcf1e71b-3d94-4450-ae39-ea10deac5dde-utilities\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.337796 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlgvk\" (UniqueName: \"kubernetes.io/projected/dcf1e71b-3d94-4450-ae39-ea10deac5dde-kube-api-access-wlgvk\") pod \"redhat-operators-q9598\" (UID: \"dcf1e71b-3d94-4450-ae39-ea10deac5dde\") " pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.558035 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.769549 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9598"] Oct 10 14:49:01 crc kubenswrapper[4788]: W1010 14:49:01.779406 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcf1e71b_3d94_4450_ae39_ea10deac5dde.slice/crio-515e17e13c9f701e80f1cdb20e4844cf93de041d300e589a80d79d20f6fa1753 WatchSource:0}: Error finding container 515e17e13c9f701e80f1cdb20e4844cf93de041d300e589a80d79d20f6fa1753: Status 404 returned error can't find the container with id 515e17e13c9f701e80f1cdb20e4844cf93de041d300e589a80d79d20f6fa1753 Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.883336 4788 generic.go:334] "Generic (PLEG): container finished" podID="f4845d06-1130-4527-a881-c01fbec64e87" containerID="ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489" exitCode=0 Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.883393 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz6z2" event={"ID":"f4845d06-1130-4527-a881-c01fbec64e87","Type":"ContainerDied","Data":"ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489"} Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.883677 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz6z2" event={"ID":"f4845d06-1130-4527-a881-c01fbec64e87","Type":"ContainerStarted","Data":"ada4a0ead8ff1ef4be4a486bd0be518aa493a05c55b05e5da85f59321db08782"} Oct 10 14:49:01 crc kubenswrapper[4788]: I1010 14:49:01.884521 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9598" event={"ID":"dcf1e71b-3d94-4450-ae39-ea10deac5dde","Type":"ContainerStarted","Data":"515e17e13c9f701e80f1cdb20e4844cf93de041d300e589a80d79d20f6fa1753"} Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.550513 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fmcvh"] Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.552074 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.555755 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.562761 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fmcvh"] Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.636962 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-utilities\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.637401 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g78z\" (UniqueName: \"kubernetes.io/projected/c86646c0-727e-4893-9d74-888fd96d2fa7-kube-api-access-9g78z\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.637493 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-catalog-content\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.740115 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-catalog-content\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.740323 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-utilities\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.740369 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g78z\" (UniqueName: \"kubernetes.io/projected/c86646c0-727e-4893-9d74-888fd96d2fa7-kube-api-access-9g78z\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.741387 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-catalog-content\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.744133 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-utilities\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.774824 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g78z\" (UniqueName: \"kubernetes.io/projected/c86646c0-727e-4893-9d74-888fd96d2fa7-kube-api-access-9g78z\") pod \"community-operators-fmcvh\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.896802 4788 generic.go:334] "Generic (PLEG): container finished" podID="dcf1e71b-3d94-4450-ae39-ea10deac5dde" containerID="cf49566d8aba833afd9a33bd7a2e4f38defa574caa3cd6aced1fc48e5a7623a0" exitCode=0 Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.896865 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9598" event={"ID":"dcf1e71b-3d94-4450-ae39-ea10deac5dde","Type":"ContainerDied","Data":"cf49566d8aba833afd9a33bd7a2e4f38defa574caa3cd6aced1fc48e5a7623a0"} Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.900761 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz6z2" event={"ID":"f4845d06-1130-4527-a881-c01fbec64e87","Type":"ContainerStarted","Data":"06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0"} Oct 10 14:49:02 crc kubenswrapper[4788]: I1010 14:49:02.903103 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.115192 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fmcvh"] Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.566072 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rskph"] Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.573016 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.585224 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.596288 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rskph"] Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.655313 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d1a4d9-5c27-447a-922e-73c4edac912c-catalog-content\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.655421 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h56z8\" (UniqueName: \"kubernetes.io/projected/85d1a4d9-5c27-447a-922e-73c4edac912c-kube-api-access-h56z8\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.655457 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d1a4d9-5c27-447a-922e-73c4edac912c-utilities\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.757375 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d1a4d9-5c27-447a-922e-73c4edac912c-catalog-content\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.757479 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h56z8\" (UniqueName: \"kubernetes.io/projected/85d1a4d9-5c27-447a-922e-73c4edac912c-kube-api-access-h56z8\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.757530 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d1a4d9-5c27-447a-922e-73c4edac912c-utilities\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.758716 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85d1a4d9-5c27-447a-922e-73c4edac912c-utilities\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.758854 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85d1a4d9-5c27-447a-922e-73c4edac912c-catalog-content\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.789561 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h56z8\" (UniqueName: \"kubernetes.io/projected/85d1a4d9-5c27-447a-922e-73c4edac912c-kube-api-access-h56z8\") pod \"redhat-marketplace-rskph\" (UID: \"85d1a4d9-5c27-447a-922e-73c4edac912c\") " pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.892234 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.909168 4788 generic.go:334] "Generic (PLEG): container finished" podID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerID="d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8" exitCode=0 Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.909278 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcvh" event={"ID":"c86646c0-727e-4893-9d74-888fd96d2fa7","Type":"ContainerDied","Data":"d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8"} Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.909347 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcvh" event={"ID":"c86646c0-727e-4893-9d74-888fd96d2fa7","Type":"ContainerStarted","Data":"7b9220823dc8130a1b8cee697c7cae8433198fa0c714500e540881b7e8156d27"} Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.913630 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9598" event={"ID":"dcf1e71b-3d94-4450-ae39-ea10deac5dde","Type":"ContainerStarted","Data":"3d74ef2b35f05761175dfbc7e7fae345c1efa0dc88fbfad05826722ab814741f"} Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.915908 4788 generic.go:334] "Generic (PLEG): container finished" podID="f4845d06-1130-4527-a881-c01fbec64e87" containerID="06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0" exitCode=0 Oct 10 14:49:03 crc kubenswrapper[4788]: I1010 14:49:03.916218 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz6z2" event={"ID":"f4845d06-1130-4527-a881-c01fbec64e87","Type":"ContainerDied","Data":"06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0"} Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.316801 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rskph"] Oct 10 14:49:04 crc kubenswrapper[4788]: W1010 14:49:04.384785 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85d1a4d9_5c27_447a_922e_73c4edac912c.slice/crio-5c6041365c79aa63750bc966a82fd3a8725a1c34bee9713b4ece690d21481aed WatchSource:0}: Error finding container 5c6041365c79aa63750bc966a82fd3a8725a1c34bee9713b4ece690d21481aed: Status 404 returned error can't find the container with id 5c6041365c79aa63750bc966a82fd3a8725a1c34bee9713b4ece690d21481aed Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.924857 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcvh" event={"ID":"c86646c0-727e-4893-9d74-888fd96d2fa7","Type":"ContainerStarted","Data":"acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9"} Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.927088 4788 generic.go:334] "Generic (PLEG): container finished" podID="85d1a4d9-5c27-447a-922e-73c4edac912c" containerID="83970abd3e6fce3080dd4c6f878aab900b27b1751b7290199af18ca4a6453f36" exitCode=0 Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.927197 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rskph" event={"ID":"85d1a4d9-5c27-447a-922e-73c4edac912c","Type":"ContainerDied","Data":"83970abd3e6fce3080dd4c6f878aab900b27b1751b7290199af18ca4a6453f36"} Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.927300 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rskph" event={"ID":"85d1a4d9-5c27-447a-922e-73c4edac912c","Type":"ContainerStarted","Data":"5c6041365c79aa63750bc966a82fd3a8725a1c34bee9713b4ece690d21481aed"} Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.929865 4788 generic.go:334] "Generic (PLEG): container finished" podID="dcf1e71b-3d94-4450-ae39-ea10deac5dde" containerID="3d74ef2b35f05761175dfbc7e7fae345c1efa0dc88fbfad05826722ab814741f" exitCode=0 Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.929944 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9598" event={"ID":"dcf1e71b-3d94-4450-ae39-ea10deac5dde","Type":"ContainerDied","Data":"3d74ef2b35f05761175dfbc7e7fae345c1efa0dc88fbfad05826722ab814741f"} Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.940205 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz6z2" event={"ID":"f4845d06-1130-4527-a881-c01fbec64e87","Type":"ContainerStarted","Data":"927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252"} Oct 10 14:49:04 crc kubenswrapper[4788]: I1010 14:49:04.980306 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lz6z2" podStartSLOduration=2.276385298 podStartE2EDuration="4.98026903s" podCreationTimestamp="2025-10-10 14:49:00 +0000 UTC" firstStartedPulling="2025-10-10 14:49:01.885081903 +0000 UTC m=+244.334797451" lastFinishedPulling="2025-10-10 14:49:04.588965615 +0000 UTC m=+247.038681183" observedRunningTime="2025-10-10 14:49:04.978962098 +0000 UTC m=+247.428677646" watchObservedRunningTime="2025-10-10 14:49:04.98026903 +0000 UTC m=+247.429984578" Oct 10 14:49:05 crc kubenswrapper[4788]: I1010 14:49:05.946860 4788 generic.go:334] "Generic (PLEG): container finished" podID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerID="acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9" exitCode=0 Oct 10 14:49:05 crc kubenswrapper[4788]: I1010 14:49:05.946985 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcvh" event={"ID":"c86646c0-727e-4893-9d74-888fd96d2fa7","Type":"ContainerDied","Data":"acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9"} Oct 10 14:49:06 crc kubenswrapper[4788]: I1010 14:49:06.956351 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcvh" event={"ID":"c86646c0-727e-4893-9d74-888fd96d2fa7","Type":"ContainerStarted","Data":"34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803"} Oct 10 14:49:06 crc kubenswrapper[4788]: I1010 14:49:06.959089 4788 generic.go:334] "Generic (PLEG): container finished" podID="85d1a4d9-5c27-447a-922e-73c4edac912c" containerID="f921ee72d5fd4fbc64e2218f34dcc19e1ef54918ad154e48c8d4261cc31d861b" exitCode=0 Oct 10 14:49:06 crc kubenswrapper[4788]: I1010 14:49:06.959187 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rskph" event={"ID":"85d1a4d9-5c27-447a-922e-73c4edac912c","Type":"ContainerDied","Data":"f921ee72d5fd4fbc64e2218f34dcc19e1ef54918ad154e48c8d4261cc31d861b"} Oct 10 14:49:06 crc kubenswrapper[4788]: I1010 14:49:06.963032 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9598" event={"ID":"dcf1e71b-3d94-4450-ae39-ea10deac5dde","Type":"ContainerStarted","Data":"85bbbe3a55fd264d1d92b0ea64928593860c06bf701cf9a25e98cf6f076830bd"} Oct 10 14:49:06 crc kubenswrapper[4788]: I1010 14:49:06.980937 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fmcvh" podStartSLOduration=2.472872715 podStartE2EDuration="4.98091494s" podCreationTimestamp="2025-10-10 14:49:02 +0000 UTC" firstStartedPulling="2025-10-10 14:49:03.910514635 +0000 UTC m=+246.360230183" lastFinishedPulling="2025-10-10 14:49:06.418556859 +0000 UTC m=+248.868272408" observedRunningTime="2025-10-10 14:49:06.976628624 +0000 UTC m=+249.426344193" watchObservedRunningTime="2025-10-10 14:49:06.98091494 +0000 UTC m=+249.430630488" Oct 10 14:49:07 crc kubenswrapper[4788]: I1010 14:49:07.001353 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q9598" podStartSLOduration=3.5585755629999998 podStartE2EDuration="6.001323155s" podCreationTimestamp="2025-10-10 14:49:01 +0000 UTC" firstStartedPulling="2025-10-10 14:49:02.89947954 +0000 UTC m=+245.349195088" lastFinishedPulling="2025-10-10 14:49:05.342227132 +0000 UTC m=+247.791942680" observedRunningTime="2025-10-10 14:49:06.994601049 +0000 UTC m=+249.444316607" watchObservedRunningTime="2025-10-10 14:49:07.001323155 +0000 UTC m=+249.451038693" Oct 10 14:49:07 crc kubenswrapper[4788]: I1010 14:49:07.972322 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rskph" event={"ID":"85d1a4d9-5c27-447a-922e-73c4edac912c","Type":"ContainerStarted","Data":"9989d1a0127d1388567f4a73e5d44216fe6aa77fbbf53d80592f973c023145c7"} Oct 10 14:49:07 crc kubenswrapper[4788]: I1010 14:49:07.997335 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rskph" podStartSLOduration=2.2713323499999998 podStartE2EDuration="4.997318398s" podCreationTimestamp="2025-10-10 14:49:03 +0000 UTC" firstStartedPulling="2025-10-10 14:49:04.928921782 +0000 UTC m=+247.378637330" lastFinishedPulling="2025-10-10 14:49:07.65490783 +0000 UTC m=+250.104623378" observedRunningTime="2025-10-10 14:49:07.995995375 +0000 UTC m=+250.445710923" watchObservedRunningTime="2025-10-10 14:49:07.997318398 +0000 UTC m=+250.447033946" Oct 10 14:49:10 crc kubenswrapper[4788]: I1010 14:49:10.512642 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:10 crc kubenswrapper[4788]: I1010 14:49:10.513171 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:10 crc kubenswrapper[4788]: I1010 14:49:10.584022 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:11 crc kubenswrapper[4788]: I1010 14:49:11.036211 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 14:49:11 crc kubenswrapper[4788]: I1010 14:49:11.558915 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:11 crc kubenswrapper[4788]: I1010 14:49:11.559254 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:11 crc kubenswrapper[4788]: I1010 14:49:11.607549 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:12 crc kubenswrapper[4788]: I1010 14:49:12.036392 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q9598" Oct 10 14:49:12 crc kubenswrapper[4788]: I1010 14:49:12.903660 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:12 crc kubenswrapper[4788]: I1010 14:49:12.903755 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:12 crc kubenswrapper[4788]: I1010 14:49:12.946509 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:13 crc kubenswrapper[4788]: I1010 14:49:13.051041 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fmcvh" Oct 10 14:49:13 crc kubenswrapper[4788]: I1010 14:49:13.893418 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:13 crc kubenswrapper[4788]: I1010 14:49:13.893540 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:13 crc kubenswrapper[4788]: I1010 14:49:13.936454 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:49:14 crc kubenswrapper[4788]: I1010 14:49:14.044544 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rskph" Oct 10 14:50:29 crc kubenswrapper[4788]: I1010 14:50:29.407128 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:50:29 crc kubenswrapper[4788]: I1010 14:50:29.408118 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:50:59 crc kubenswrapper[4788]: I1010 14:50:59.405895 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:50:59 crc kubenswrapper[4788]: I1010 14:50:59.406475 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.406078 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.406804 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.406873 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.407764 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cc73d4024eb0529c2e8485e4f790187ca0d08a7d42f2bdd40bacb41d592bece3"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.407872 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://cc73d4024eb0529c2e8485e4f790187ca0d08a7d42f2bdd40bacb41d592bece3" gracePeriod=600 Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.908224 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="cc73d4024eb0529c2e8485e4f790187ca0d08a7d42f2bdd40bacb41d592bece3" exitCode=0 Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.908302 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"cc73d4024eb0529c2e8485e4f790187ca0d08a7d42f2bdd40bacb41d592bece3"} Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.908642 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"25ef4cac1d4334ffa148701a58072ef08a1a55c6fbfe78f6ba004da25cafdd99"} Oct 10 14:51:29 crc kubenswrapper[4788]: I1010 14:51:29.908669 4788 scope.go:117] "RemoveContainer" containerID="831ecb56df69be1ceb0a45cb1f32b152df855fb95a1c86259c7221bf319ab817" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.658234 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-glv4r"] Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.659718 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.677874 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-glv4r"] Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.766402 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-registry-tls\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.766880 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpxmx\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-kube-api-access-fpxmx\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.766914 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-bound-sa-token\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.766931 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c74a7a2-216a-4346-8bb4-997812df8155-ca-trust-extracted\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.766956 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c74a7a2-216a-4346-8bb4-997812df8155-trusted-ca\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.766980 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c74a7a2-216a-4346-8bb4-997812df8155-installation-pull-secrets\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.767156 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c74a7a2-216a-4346-8bb4-997812df8155-registry-certificates\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.767309 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.793271 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.868426 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-registry-tls\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.868489 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpxmx\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-kube-api-access-fpxmx\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.868511 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-bound-sa-token\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.868527 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c74a7a2-216a-4346-8bb4-997812df8155-ca-trust-extracted\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.868553 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c74a7a2-216a-4346-8bb4-997812df8155-trusted-ca\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.868571 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c74a7a2-216a-4346-8bb4-997812df8155-installation-pull-secrets\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.868597 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c74a7a2-216a-4346-8bb4-997812df8155-registry-certificates\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.869642 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8c74a7a2-216a-4346-8bb4-997812df8155-ca-trust-extracted\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.870272 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8c74a7a2-216a-4346-8bb4-997812df8155-registry-certificates\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.870424 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c74a7a2-216a-4346-8bb4-997812df8155-trusted-ca\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.876595 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-registry-tls\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.878880 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8c74a7a2-216a-4346-8bb4-997812df8155-installation-pull-secrets\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.888016 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-bound-sa-token\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.888460 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpxmx\" (UniqueName: \"kubernetes.io/projected/8c74a7a2-216a-4346-8bb4-997812df8155-kube-api-access-fpxmx\") pod \"image-registry-66df7c8f76-glv4r\" (UID: \"8c74a7a2-216a-4346-8bb4-997812df8155\") " pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:23 crc kubenswrapper[4788]: I1010 14:53:23.974107 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:24 crc kubenswrapper[4788]: I1010 14:53:24.174647 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-glv4r"] Oct 10 14:53:24 crc kubenswrapper[4788]: I1010 14:53:24.728775 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" event={"ID":"8c74a7a2-216a-4346-8bb4-997812df8155","Type":"ContainerStarted","Data":"abae7859ce51008c60eb242dc2f0df57308a9ded56b9e2dd05539017886efa63"} Oct 10 14:53:24 crc kubenswrapper[4788]: I1010 14:53:24.729241 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" event={"ID":"8c74a7a2-216a-4346-8bb4-997812df8155","Type":"ContainerStarted","Data":"7786cd15c773586db254561885927bd1777789fe42313c902e86fd96bbbf0ea9"} Oct 10 14:53:24 crc kubenswrapper[4788]: I1010 14:53:24.729264 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:29 crc kubenswrapper[4788]: I1010 14:53:29.408271 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:53:29 crc kubenswrapper[4788]: I1010 14:53:29.408865 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:53:43 crc kubenswrapper[4788]: I1010 14:53:43.982013 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" Oct 10 14:53:44 crc kubenswrapper[4788]: I1010 14:53:44.005888 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-glv4r" podStartSLOduration=21.005864747 podStartE2EDuration="21.005864747s" podCreationTimestamp="2025-10-10 14:53:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:53:24.761130681 +0000 UTC m=+507.210846229" watchObservedRunningTime="2025-10-10 14:53:44.005864747 +0000 UTC m=+526.455580315" Oct 10 14:53:44 crc kubenswrapper[4788]: I1010 14:53:44.045116 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kdbvl"] Oct 10 14:53:58 crc kubenswrapper[4788]: I1010 14:53:58.441072 4788 scope.go:117] "RemoveContainer" containerID="08066560a1488d814ab12d68ba5b0d935e8aa03a30e435c6570b213239ef8840" Oct 10 14:53:58 crc kubenswrapper[4788]: I1010 14:53:58.466528 4788 scope.go:117] "RemoveContainer" containerID="5fe194d114f06da7c83ea5e21969f4eec3793d43a44602612ea8643a5115ab3b" Oct 10 14:53:59 crc kubenswrapper[4788]: I1010 14:53:59.406817 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:53:59 crc kubenswrapper[4788]: I1010 14:53:59.406928 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.093379 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" podUID="2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" containerName="registry" containerID="cri-o://fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8" gracePeriod=30 Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.480901 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.615678 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.615793 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-tls\") pod \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.615865 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-trusted-ca\") pod \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.615923 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrqj5\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-kube-api-access-wrqj5\") pod \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.615965 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-ca-trust-extracted\") pod \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.615999 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-installation-pull-secrets\") pod \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.616033 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-certificates\") pod \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.616071 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-bound-sa-token\") pod \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\" (UID: \"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce\") " Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.617444 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.617803 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.620505 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.624336 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.624862 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-kube-api-access-wrqj5" (OuterVolumeSpecName: "kube-api-access-wrqj5") pod "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce"). InnerVolumeSpecName "kube-api-access-wrqj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.627257 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.627455 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.628765 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.650163 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" (UID: "2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.719358 4788 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.719404 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrqj5\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-kube-api-access-wrqj5\") on node \"crc\" DevicePath \"\"" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.719420 4788 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.719433 4788 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.719446 4788 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 10 14:54:09 crc kubenswrapper[4788]: I1010 14:54:09.719459 4788 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.006292 4788 generic.go:334] "Generic (PLEG): container finished" podID="2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" containerID="fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8" exitCode=0 Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.006360 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.006385 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" event={"ID":"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce","Type":"ContainerDied","Data":"fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8"} Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.007213 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kdbvl" event={"ID":"2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce","Type":"ContainerDied","Data":"48e63761b0c12c77c6717431c9a535c4b6529012d464fc08970afe59064efcb3"} Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.007264 4788 scope.go:117] "RemoveContainer" containerID="fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8" Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.026756 4788 scope.go:117] "RemoveContainer" containerID="fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8" Oct 10 14:54:10 crc kubenswrapper[4788]: E1010 14:54:10.027545 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8\": container with ID starting with fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8 not found: ID does not exist" containerID="fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8" Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.027582 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8"} err="failed to get container status \"fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8\": rpc error: code = NotFound desc = could not find container \"fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8\": container with ID starting with fe01a383f19971441a26a225afad2eba2e581762586e94678ca1f4a3081cefc8 not found: ID does not exist" Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.055534 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kdbvl"] Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.060056 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kdbvl"] Oct 10 14:54:10 crc kubenswrapper[4788]: I1010 14:54:10.246782 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" path="/var/lib/kubelet/pods/2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce/volumes" Oct 10 14:54:29 crc kubenswrapper[4788]: I1010 14:54:29.406838 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:54:29 crc kubenswrapper[4788]: I1010 14:54:29.407935 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:54:29 crc kubenswrapper[4788]: I1010 14:54:29.408031 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:54:29 crc kubenswrapper[4788]: I1010 14:54:29.409135 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25ef4cac1d4334ffa148701a58072ef08a1a55c6fbfe78f6ba004da25cafdd99"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:54:29 crc kubenswrapper[4788]: I1010 14:54:29.409300 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://25ef4cac1d4334ffa148701a58072ef08a1a55c6fbfe78f6ba004da25cafdd99" gracePeriod=600 Oct 10 14:54:30 crc kubenswrapper[4788]: I1010 14:54:30.154749 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="25ef4cac1d4334ffa148701a58072ef08a1a55c6fbfe78f6ba004da25cafdd99" exitCode=0 Oct 10 14:54:30 crc kubenswrapper[4788]: I1010 14:54:30.154900 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"25ef4cac1d4334ffa148701a58072ef08a1a55c6fbfe78f6ba004da25cafdd99"} Oct 10 14:54:30 crc kubenswrapper[4788]: I1010 14:54:30.155356 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"10d046d777ded5357953f35d670fff38bbf3941e6d3d9f85f546d304830131b6"} Oct 10 14:54:30 crc kubenswrapper[4788]: I1010 14:54:30.155408 4788 scope.go:117] "RemoveContainer" containerID="cc73d4024eb0529c2e8485e4f790187ca0d08a7d42f2bdd40bacb41d592bece3" Oct 10 14:54:58 crc kubenswrapper[4788]: I1010 14:54:58.517924 4788 scope.go:117] "RemoveContainer" containerID="824767b6041c1c6836f4a45fe6c5cdaf3292508af0046fed3a23a104f78f153e" Oct 10 14:56:29 crc kubenswrapper[4788]: I1010 14:56:29.407071 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:56:29 crc kubenswrapper[4788]: I1010 14:56:29.408055 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:56:59 crc kubenswrapper[4788]: I1010 14:56:59.406210 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:56:59 crc kubenswrapper[4788]: I1010 14:56:59.409052 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:57:24 crc kubenswrapper[4788]: I1010 14:57:24.710859 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m448m"] Oct 10 14:57:24 crc kubenswrapper[4788]: I1010 14:57:24.712462 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" podUID="10e304aa-8568-440e-99cc-f9081412645e" containerName="controller-manager" containerID="cri-o://bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a" gracePeriod=30 Oct 10 14:57:24 crc kubenswrapper[4788]: I1010 14:57:24.792275 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq"] Oct 10 14:57:24 crc kubenswrapper[4788]: I1010 14:57:24.792592 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" podUID="51751bfc-73b2-4968-bbf7-a213e109b40a" containerName="route-controller-manager" containerID="cri-o://f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198" gracePeriod=30 Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.211066 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.217030 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.312707 4788 generic.go:334] "Generic (PLEG): container finished" podID="10e304aa-8568-440e-99cc-f9081412645e" containerID="bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a" exitCode=0 Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.312787 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" event={"ID":"10e304aa-8568-440e-99cc-f9081412645e","Type":"ContainerDied","Data":"bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a"} Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.312820 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" event={"ID":"10e304aa-8568-440e-99cc-f9081412645e","Type":"ContainerDied","Data":"909f1e8746ae0b2a5cc01fcc05e5ac042ff869f55a13324be8167ad86aab8bbd"} Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.312818 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m448m" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.312839 4788 scope.go:117] "RemoveContainer" containerID="bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.315071 4788 generic.go:334] "Generic (PLEG): container finished" podID="51751bfc-73b2-4968-bbf7-a213e109b40a" containerID="f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198" exitCode=0 Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.315118 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.315137 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" event={"ID":"51751bfc-73b2-4968-bbf7-a213e109b40a","Type":"ContainerDied","Data":"f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198"} Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.315216 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq" event={"ID":"51751bfc-73b2-4968-bbf7-a213e109b40a","Type":"ContainerDied","Data":"ab33f62f4ff25bd46501cc6f50e1f4b8c8e6a72617bbb6957f253160d1ee138c"} Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.345089 4788 scope.go:117] "RemoveContainer" containerID="bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a" Oct 10 14:57:25 crc kubenswrapper[4788]: E1010 14:57:25.345696 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a\": container with ID starting with bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a not found: ID does not exist" containerID="bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.345738 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a"} err="failed to get container status \"bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a\": rpc error: code = NotFound desc = could not find container \"bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a\": container with ID starting with bd13e7445f88c6c63b948fbfa61e5fdef14910f1bf9a5adf3f5a120495f92a2a not found: ID does not exist" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.345767 4788 scope.go:117] "RemoveContainer" containerID="f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.382773 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbjt9\" (UniqueName: \"kubernetes.io/projected/10e304aa-8568-440e-99cc-f9081412645e-kube-api-access-qbjt9\") pod \"10e304aa-8568-440e-99cc-f9081412645e\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.382883 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e304aa-8568-440e-99cc-f9081412645e-serving-cert\") pod \"10e304aa-8568-440e-99cc-f9081412645e\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.382920 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-config\") pod \"10e304aa-8568-440e-99cc-f9081412645e\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.382974 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51751bfc-73b2-4968-bbf7-a213e109b40a-serving-cert\") pod \"51751bfc-73b2-4968-bbf7-a213e109b40a\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.383018 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-client-ca\") pod \"10e304aa-8568-440e-99cc-f9081412645e\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.383050 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-proxy-ca-bundles\") pod \"10e304aa-8568-440e-99cc-f9081412645e\" (UID: \"10e304aa-8568-440e-99cc-f9081412645e\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.383184 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-config\") pod \"51751bfc-73b2-4968-bbf7-a213e109b40a\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.383226 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-client-ca\") pod \"51751bfc-73b2-4968-bbf7-a213e109b40a\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.383252 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nt68\" (UniqueName: \"kubernetes.io/projected/51751bfc-73b2-4968-bbf7-a213e109b40a-kube-api-access-6nt68\") pod \"51751bfc-73b2-4968-bbf7-a213e109b40a\" (UID: \"51751bfc-73b2-4968-bbf7-a213e109b40a\") " Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.384421 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-client-ca" (OuterVolumeSpecName: "client-ca") pod "10e304aa-8568-440e-99cc-f9081412645e" (UID: "10e304aa-8568-440e-99cc-f9081412645e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.384971 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "10e304aa-8568-440e-99cc-f9081412645e" (UID: "10e304aa-8568-440e-99cc-f9081412645e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.385013 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-config" (OuterVolumeSpecName: "config") pod "10e304aa-8568-440e-99cc-f9081412645e" (UID: "10e304aa-8568-440e-99cc-f9081412645e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.385042 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-client-ca" (OuterVolumeSpecName: "client-ca") pod "51751bfc-73b2-4968-bbf7-a213e109b40a" (UID: "51751bfc-73b2-4968-bbf7-a213e109b40a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.385368 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-config" (OuterVolumeSpecName: "config") pod "51751bfc-73b2-4968-bbf7-a213e109b40a" (UID: "51751bfc-73b2-4968-bbf7-a213e109b40a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.390852 4788 scope.go:117] "RemoveContainer" containerID="f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.391716 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51751bfc-73b2-4968-bbf7-a213e109b40a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "51751bfc-73b2-4968-bbf7-a213e109b40a" (UID: "51751bfc-73b2-4968-bbf7-a213e109b40a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: E1010 14:57:25.391853 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198\": container with ID starting with f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198 not found: ID does not exist" containerID="f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.391896 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198"} err="failed to get container status \"f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198\": rpc error: code = NotFound desc = could not find container \"f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198\": container with ID starting with f3b26c7d1c109ab866a50b887036031ed3737a9b3f9fa5f9ac7e39b90d468198 not found: ID does not exist" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.392012 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10e304aa-8568-440e-99cc-f9081412645e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "10e304aa-8568-440e-99cc-f9081412645e" (UID: "10e304aa-8568-440e-99cc-f9081412645e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.392127 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10e304aa-8568-440e-99cc-f9081412645e-kube-api-access-qbjt9" (OuterVolumeSpecName: "kube-api-access-qbjt9") pod "10e304aa-8568-440e-99cc-f9081412645e" (UID: "10e304aa-8568-440e-99cc-f9081412645e"). InnerVolumeSpecName "kube-api-access-qbjt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.392681 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51751bfc-73b2-4968-bbf7-a213e109b40a-kube-api-access-6nt68" (OuterVolumeSpecName: "kube-api-access-6nt68") pod "51751bfc-73b2-4968-bbf7-a213e109b40a" (UID: "51751bfc-73b2-4968-bbf7-a213e109b40a"). InnerVolumeSpecName "kube-api-access-6nt68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.484926 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51751bfc-73b2-4968-bbf7-a213e109b40a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.484968 4788 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.484978 4788 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.484987 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.484995 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nt68\" (UniqueName: \"kubernetes.io/projected/51751bfc-73b2-4968-bbf7-a213e109b40a-kube-api-access-6nt68\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.485007 4788 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51751bfc-73b2-4968-bbf7-a213e109b40a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.485015 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbjt9\" (UniqueName: \"kubernetes.io/projected/10e304aa-8568-440e-99cc-f9081412645e-kube-api-access-qbjt9\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.485023 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e304aa-8568-440e-99cc-f9081412645e-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.485034 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e304aa-8568-440e-99cc-f9081412645e-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.652103 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m448m"] Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.660370 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m448m"] Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.675288 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq"] Oct 10 14:57:25 crc kubenswrapper[4788]: I1010 14:57:25.680789 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rnqfq"] Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.152407 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7644ddbf8d-v86cq"] Oct 10 14:57:26 crc kubenswrapper[4788]: E1010 14:57:26.153941 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10e304aa-8568-440e-99cc-f9081412645e" containerName="controller-manager" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.154035 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="10e304aa-8568-440e-99cc-f9081412645e" containerName="controller-manager" Oct 10 14:57:26 crc kubenswrapper[4788]: E1010 14:57:26.154168 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" containerName="registry" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.154240 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" containerName="registry" Oct 10 14:57:26 crc kubenswrapper[4788]: E1010 14:57:26.154346 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51751bfc-73b2-4968-bbf7-a213e109b40a" containerName="route-controller-manager" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.154448 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="51751bfc-73b2-4968-bbf7-a213e109b40a" containerName="route-controller-manager" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.154638 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="51751bfc-73b2-4968-bbf7-a213e109b40a" containerName="route-controller-manager" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.154715 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c1c0b3c-a8ed-46bd-8f7f-beaf6c607fce" containerName="registry" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.154797 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="10e304aa-8568-440e-99cc-f9081412645e" containerName="controller-manager" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.155459 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.158116 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2"] Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.159339 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.159490 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.160208 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.160911 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.161201 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.161985 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.162003 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.164309 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.164957 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.165008 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.165455 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.172624 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.173972 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.178873 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.179629 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2"] Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.185198 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7644ddbf8d-v86cq"] Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.247048 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10e304aa-8568-440e-99cc-f9081412645e" path="/var/lib/kubelet/pods/10e304aa-8568-440e-99cc-f9081412645e/volumes" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.248790 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51751bfc-73b2-4968-bbf7-a213e109b40a" path="/var/lib/kubelet/pods/51751bfc-73b2-4968-bbf7-a213e109b40a/volumes" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.297659 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-client-ca\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.297780 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-client-ca\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.297850 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-config\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.298080 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-proxy-ca-bundles\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.298241 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/487c1deb-6b68-4ffd-ab67-6b77e8747e20-serving-cert\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.298812 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-serving-cert\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.298880 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-config\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.298927 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh2bs\" (UniqueName: \"kubernetes.io/projected/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-kube-api-access-dh2bs\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.298968 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24cpn\" (UniqueName: \"kubernetes.io/projected/487c1deb-6b68-4ffd-ab67-6b77e8747e20-kube-api-access-24cpn\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.326671 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2"] Oct 10 14:57:26 crc kubenswrapper[4788]: E1010 14:57:26.327455 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-24cpn serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" podUID="487c1deb-6b68-4ffd-ab67-6b77e8747e20" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400203 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-client-ca\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400293 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-config\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400387 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-proxy-ca-bundles\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400456 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/487c1deb-6b68-4ffd-ab67-6b77e8747e20-serving-cert\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400499 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-serving-cert\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400532 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-config\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400567 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh2bs\" (UniqueName: \"kubernetes.io/projected/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-kube-api-access-dh2bs\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400607 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24cpn\" (UniqueName: \"kubernetes.io/projected/487c1deb-6b68-4ffd-ab67-6b77e8747e20-kube-api-access-24cpn\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.400677 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-client-ca\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.401769 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-client-ca\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.402284 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-client-ca\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.402550 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-config\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.403401 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-config\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.403447 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-proxy-ca-bundles\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.408059 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/487c1deb-6b68-4ffd-ab67-6b77e8747e20-serving-cert\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.408971 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-serving-cert\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.426032 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24cpn\" (UniqueName: \"kubernetes.io/projected/487c1deb-6b68-4ffd-ab67-6b77e8747e20-kube-api-access-24cpn\") pod \"route-controller-manager-7db768bfbb-q6vp2\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.433706 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh2bs\" (UniqueName: \"kubernetes.io/projected/06b882d2-9ba2-4b9e-8ea8-7e233e173c0a-kube-api-access-dh2bs\") pod \"controller-manager-7644ddbf8d-v86cq\" (UID: \"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a\") " pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.521137 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:26 crc kubenswrapper[4788]: I1010 14:57:26.767794 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7644ddbf8d-v86cq"] Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.332880 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.332877 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" event={"ID":"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a","Type":"ContainerStarted","Data":"bf586ce5074fc20a3ae8f8e81b4df71d287a5ac61b0139bf288b59cffd086cff"} Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.333628 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.333660 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" event={"ID":"06b882d2-9ba2-4b9e-8ea8-7e233e173c0a","Type":"ContainerStarted","Data":"380fd72b1076f9649f23c728d81a9f573ec1e88201a9632d42fbca34d2dc1410"} Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.345728 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.347264 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.376157 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7644ddbf8d-v86cq" podStartSLOduration=3.37610296 podStartE2EDuration="3.37610296s" podCreationTimestamp="2025-10-10 14:57:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:57:27.354965 +0000 UTC m=+749.804680558" watchObservedRunningTime="2025-10-10 14:57:27.37610296 +0000 UTC m=+749.825818518" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.519048 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-client-ca\") pod \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.519159 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24cpn\" (UniqueName: \"kubernetes.io/projected/487c1deb-6b68-4ffd-ab67-6b77e8747e20-kube-api-access-24cpn\") pod \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.519322 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-config\") pod \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.519360 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/487c1deb-6b68-4ffd-ab67-6b77e8747e20-serving-cert\") pod \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\" (UID: \"487c1deb-6b68-4ffd-ab67-6b77e8747e20\") " Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.519880 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-client-ca" (OuterVolumeSpecName: "client-ca") pod "487c1deb-6b68-4ffd-ab67-6b77e8747e20" (UID: "487c1deb-6b68-4ffd-ab67-6b77e8747e20"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.520191 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-config" (OuterVolumeSpecName: "config") pod "487c1deb-6b68-4ffd-ab67-6b77e8747e20" (UID: "487c1deb-6b68-4ffd-ab67-6b77e8747e20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.550014 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/487c1deb-6b68-4ffd-ab67-6b77e8747e20-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "487c1deb-6b68-4ffd-ab67-6b77e8747e20" (UID: "487c1deb-6b68-4ffd-ab67-6b77e8747e20"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.550119 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/487c1deb-6b68-4ffd-ab67-6b77e8747e20-kube-api-access-24cpn" (OuterVolumeSpecName: "kube-api-access-24cpn") pod "487c1deb-6b68-4ffd-ab67-6b77e8747e20" (UID: "487c1deb-6b68-4ffd-ab67-6b77e8747e20"). InnerVolumeSpecName "kube-api-access-24cpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.621181 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.621224 4788 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/487c1deb-6b68-4ffd-ab67-6b77e8747e20-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.621234 4788 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/487c1deb-6b68-4ffd-ab67-6b77e8747e20-client-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:27 crc kubenswrapper[4788]: I1010 14:57:27.621247 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24cpn\" (UniqueName: \"kubernetes.io/projected/487c1deb-6b68-4ffd-ab67-6b77e8747e20-kube-api-access-24cpn\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.339158 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.374646 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2"] Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.384554 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7db768bfbb-q6vp2"] Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.397865 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts"] Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.399666 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.403578 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts"] Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.404510 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.404932 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.405280 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.405349 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.405756 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.406092 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.434714 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwpcf\" (UniqueName: \"kubernetes.io/projected/a005c202-e5ce-45d7-896e-05e28ada4465-kube-api-access-mwpcf\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.434778 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a005c202-e5ce-45d7-896e-05e28ada4465-config\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.434831 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a005c202-e5ce-45d7-896e-05e28ada4465-client-ca\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.434863 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a005c202-e5ce-45d7-896e-05e28ada4465-serving-cert\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.536180 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwpcf\" (UniqueName: \"kubernetes.io/projected/a005c202-e5ce-45d7-896e-05e28ada4465-kube-api-access-mwpcf\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.536278 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a005c202-e5ce-45d7-896e-05e28ada4465-config\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.536343 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a005c202-e5ce-45d7-896e-05e28ada4465-client-ca\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.536385 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a005c202-e5ce-45d7-896e-05e28ada4465-serving-cert\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.538086 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a005c202-e5ce-45d7-896e-05e28ada4465-config\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.538573 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a005c202-e5ce-45d7-896e-05e28ada4465-client-ca\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.547259 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a005c202-e5ce-45d7-896e-05e28ada4465-serving-cert\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.588190 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwpcf\" (UniqueName: \"kubernetes.io/projected/a005c202-e5ce-45d7-896e-05e28ada4465-kube-api-access-mwpcf\") pod \"route-controller-manager-6666cb94f6-kbfts\" (UID: \"a005c202-e5ce-45d7-896e-05e28ada4465\") " pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.722162 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:28 crc kubenswrapper[4788]: I1010 14:57:28.975533 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts"] Oct 10 14:57:28 crc kubenswrapper[4788]: W1010 14:57:28.988714 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda005c202_e5ce_45d7_896e_05e28ada4465.slice/crio-fce49dbb4732339e89009b2cb454128fb7d6c7a786fec09b6d120fa9d000921e WatchSource:0}: Error finding container fce49dbb4732339e89009b2cb454128fb7d6c7a786fec09b6d120fa9d000921e: Status 404 returned error can't find the container with id fce49dbb4732339e89009b2cb454128fb7d6c7a786fec09b6d120fa9d000921e Oct 10 14:57:29 crc kubenswrapper[4788]: I1010 14:57:29.347611 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" event={"ID":"a005c202-e5ce-45d7-896e-05e28ada4465","Type":"ContainerStarted","Data":"c0330f5934824c3a93d1d41b8bfaf835383b7f7c1aa23440049535eb49cc3728"} Oct 10 14:57:29 crc kubenswrapper[4788]: I1010 14:57:29.347698 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" event={"ID":"a005c202-e5ce-45d7-896e-05e28ada4465","Type":"ContainerStarted","Data":"fce49dbb4732339e89009b2cb454128fb7d6c7a786fec09b6d120fa9d000921e"} Oct 10 14:57:29 crc kubenswrapper[4788]: I1010 14:57:29.369264 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" podStartSLOduration=3.369245442 podStartE2EDuration="3.369245442s" podCreationTimestamp="2025-10-10 14:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:57:29.366927617 +0000 UTC m=+751.816643175" watchObservedRunningTime="2025-10-10 14:57:29.369245442 +0000 UTC m=+751.818960990" Oct 10 14:57:29 crc kubenswrapper[4788]: I1010 14:57:29.406564 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:57:29 crc kubenswrapper[4788]: I1010 14:57:29.406672 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:57:29 crc kubenswrapper[4788]: I1010 14:57:29.406752 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 14:57:29 crc kubenswrapper[4788]: I1010 14:57:29.407876 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"10d046d777ded5357953f35d670fff38bbf3941e6d3d9f85f546d304830131b6"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 14:57:29 crc kubenswrapper[4788]: I1010 14:57:29.408001 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://10d046d777ded5357953f35d670fff38bbf3941e6d3d9f85f546d304830131b6" gracePeriod=600 Oct 10 14:57:30 crc kubenswrapper[4788]: I1010 14:57:30.240758 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="487c1deb-6b68-4ffd-ab67-6b77e8747e20" path="/var/lib/kubelet/pods/487c1deb-6b68-4ffd-ab67-6b77e8747e20/volumes" Oct 10 14:57:30 crc kubenswrapper[4788]: I1010 14:57:30.354834 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="10d046d777ded5357953f35d670fff38bbf3941e6d3d9f85f546d304830131b6" exitCode=0 Oct 10 14:57:30 crc kubenswrapper[4788]: I1010 14:57:30.354921 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"10d046d777ded5357953f35d670fff38bbf3941e6d3d9f85f546d304830131b6"} Oct 10 14:57:30 crc kubenswrapper[4788]: I1010 14:57:30.354989 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"d62f45676977d2854ac3c581616b717de6d75ef758d15d869e78e7d7a3a145f9"} Oct 10 14:57:30 crc kubenswrapper[4788]: I1010 14:57:30.355012 4788 scope.go:117] "RemoveContainer" containerID="25ef4cac1d4334ffa148701a58072ef08a1a55c6fbfe78f6ba004da25cafdd99" Oct 10 14:57:30 crc kubenswrapper[4788]: I1010 14:57:30.355939 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:30 crc kubenswrapper[4788]: I1010 14:57:30.362082 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6666cb94f6-kbfts" Oct 10 14:57:34 crc kubenswrapper[4788]: I1010 14:57:34.633616 4788 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.381478 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cwgns"] Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.383629 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovn-controller" containerID="cri-o://a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65" gracePeriod=30 Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.384187 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="sbdb" containerID="cri-o://574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401" gracePeriod=30 Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.384241 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="nbdb" containerID="cri-o://498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3" gracePeriod=30 Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.384289 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="northd" containerID="cri-o://4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d" gracePeriod=30 Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.384331 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381" gracePeriod=30 Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.384380 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kube-rbac-proxy-node" containerID="cri-o://0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3" gracePeriod=30 Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.384427 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovn-acl-logging" containerID="cri-o://41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab" gracePeriod=30 Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.427085 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" containerID="cri-o://e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0" gracePeriod=30 Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.730070 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/3.log" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.732597 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovn-acl-logging/0.log" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.733254 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovn-controller/0.log" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.733774 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816438 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8v4sx"] Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816652 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816664 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816673 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kube-rbac-proxy-node" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816679 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kube-rbac-proxy-node" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816688 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816694 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816705 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovn-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816711 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovn-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816718 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovn-acl-logging" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816723 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovn-acl-logging" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816731 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="northd" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816737 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="northd" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816745 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kubecfg-setup" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816752 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kubecfg-setup" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816761 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816767 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816775 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="sbdb" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816781 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="sbdb" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816789 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816795 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.816802 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="nbdb" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816808 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="nbdb" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816899 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="nbdb" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816908 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovn-acl-logging" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816917 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816924 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kube-rbac-proxy-ovn-metrics" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816930 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="sbdb" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816941 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816948 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="northd" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816955 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="kube-rbac-proxy-node" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816962 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816968 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816976 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovn-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.816983 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.817075 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.817082 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: E1010 14:57:39.817091 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.817097 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" containerName="ovnkube-controller" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.818780 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.892960 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-systemd\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893043 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-netd\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893121 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-openvswitch\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893211 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-config\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893237 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893334 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893391 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-systemd-units\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893419 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893552 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-var-lib-openvswitch\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893600 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/baeca99f-afc8-4129-8110-51e79dd8b054-ovn-node-metrics-cert\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893634 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-etc-openvswitch\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893661 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-log-socket\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893681 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-ovn-kubernetes\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893713 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-script-lib\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893755 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-ovn\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893778 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-node-log\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893808 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-var-lib-cni-networks-ovn-kubernetes\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893840 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-env-overrides\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893834 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893872 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-netns\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893898 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894091 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894123 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894177 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-log-socket" (OuterVolumeSpecName: "log-socket") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.893906 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kckf4\" (UniqueName: \"kubernetes.io/projected/baeca99f-afc8-4129-8110-51e79dd8b054-kube-api-access-kckf4\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894212 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894239 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-node-log" (OuterVolumeSpecName: "node-log") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894461 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894403 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-kubelet\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894287 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894584 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-slash\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894623 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-bin\") pod \"baeca99f-afc8-4129-8110-51e79dd8b054\" (UID: \"baeca99f-afc8-4129-8110-51e79dd8b054\") " Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894728 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-slash" (OuterVolumeSpecName: "host-slash") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894753 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894936 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-systemd\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895291 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-log-socket\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.894882 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895522 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-var-lib-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895691 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-node-log\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895707 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895730 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-ovnkube-config\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895757 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895789 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-slash\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895810 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-etc-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895840 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-env-overrides\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895861 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-cni-netd\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895933 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.895962 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-systemd-units\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896028 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-kubelet\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896057 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5e62642c-2473-488a-acf5-5f6903918e4f-ovn-node-metrics-cert\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896086 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7gh5\" (UniqueName: \"kubernetes.io/projected/5e62642c-2473-488a-acf5-5f6903918e4f-kube-api-access-w7gh5\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896208 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-ovn\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896232 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-cni-bin\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896262 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896287 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-ovnkube-script-lib\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896352 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-run-netns\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896384 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-run-ovn-kubernetes\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896459 4788 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896474 4788 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896484 4788 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896497 4788 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896508 4788 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896517 4788 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896529 4788 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-log-socket\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896539 4788 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896552 4788 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896562 4788 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896574 4788 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-node-log\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896585 4788 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896597 4788 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/baeca99f-afc8-4129-8110-51e79dd8b054-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896607 4788 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896619 4788 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896695 4788 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-slash\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.896706 4788 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.899734 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baeca99f-afc8-4129-8110-51e79dd8b054-kube-api-access-kckf4" (OuterVolumeSpecName: "kube-api-access-kckf4") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "kube-api-access-kckf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.900024 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baeca99f-afc8-4129-8110-51e79dd8b054-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.907225 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "baeca99f-afc8-4129-8110-51e79dd8b054" (UID: "baeca99f-afc8-4129-8110-51e79dd8b054"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998089 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998215 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-slash\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998253 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-etc-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998301 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-env-overrides\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998337 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-cni-netd\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998389 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-systemd-units\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998429 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-kubelet\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998466 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5e62642c-2473-488a-acf5-5f6903918e4f-ovn-node-metrics-cert\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998498 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7gh5\" (UniqueName: \"kubernetes.io/projected/5e62642c-2473-488a-acf5-5f6903918e4f-kube-api-access-w7gh5\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998531 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998563 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-cni-bin\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998632 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-ovn\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998640 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-cni-bin\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998661 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998691 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998709 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-ovnkube-script-lib\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998730 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-slash\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998746 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-run-netns\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998761 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-ovn\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998784 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-run-ovn-kubernetes\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998824 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-systemd\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998856 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-log-socket\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998897 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-var-lib-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998936 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-node-log\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.998966 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-ovnkube-config\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.999033 4788 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/baeca99f-afc8-4129-8110-51e79dd8b054-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.999056 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/baeca99f-afc8-4129-8110-51e79dd8b054-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.999079 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kckf4\" (UniqueName: \"kubernetes.io/projected/baeca99f-afc8-4129-8110-51e79dd8b054-kube-api-access-kckf4\") on node \"crc\" DevicePath \"\"" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.999668 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-cni-netd\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:39 crc kubenswrapper[4788]: I1010 14:57:39.999787 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-run-ovn-kubernetes\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.000054 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-run-netns\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:39.998490 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-etc-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.000208 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-var-lib-openvswitch\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.000245 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-node-log\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.000236 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-log-socket\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.000300 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-systemd-units\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.000293 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-host-kubelet\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.000272 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5e62642c-2473-488a-acf5-5f6903918e4f-run-systemd\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.001241 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-ovnkube-config\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.001978 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-env-overrides\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.002964 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5e62642c-2473-488a-acf5-5f6903918e4f-ovnkube-script-lib\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.003412 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5e62642c-2473-488a-acf5-5f6903918e4f-ovn-node-metrics-cert\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.016248 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7gh5\" (UniqueName: \"kubernetes.io/projected/5e62642c-2473-488a-acf5-5f6903918e4f-kube-api-access-w7gh5\") pod \"ovnkube-node-8v4sx\" (UID: \"5e62642c-2473-488a-acf5-5f6903918e4f\") " pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.134745 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:40 crc kubenswrapper[4788]: W1010 14:57:40.155954 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e62642c_2473_488a_acf5_5f6903918e4f.slice/crio-65a5820e72c3c58e4fe0ea106400e3fc62a8b87769379b710a83271c96bb8690 WatchSource:0}: Error finding container 65a5820e72c3c58e4fe0ea106400e3fc62a8b87769379b710a83271c96bb8690: Status 404 returned error can't find the container with id 65a5820e72c3c58e4fe0ea106400e3fc62a8b87769379b710a83271c96bb8690 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.415556 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/2.log" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.416642 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/1.log" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.416690 4788 generic.go:334] "Generic (PLEG): container finished" podID="39cfa1ec-4912-4550-9ead-7f6113db2221" containerID="e12ad039a71ffba781a71f8c01ee60626369df48ce03e2018de3b6a2dfc16c0f" exitCode=2 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.416732 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rzgfk" event={"ID":"39cfa1ec-4912-4550-9ead-7f6113db2221","Type":"ContainerDied","Data":"e12ad039a71ffba781a71f8c01ee60626369df48ce03e2018de3b6a2dfc16c0f"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.416784 4788 scope.go:117] "RemoveContainer" containerID="a38a4d328900ae940a07101123be2fc5233fec7126ab5c59b8dbb68bf87427be" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.417941 4788 scope.go:117] "RemoveContainer" containerID="e12ad039a71ffba781a71f8c01ee60626369df48ce03e2018de3b6a2dfc16c0f" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.419637 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovnkube-controller/3.log" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.429740 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovn-acl-logging/0.log" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.430794 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cwgns_baeca99f-afc8-4129-8110-51e79dd8b054/ovn-controller/0.log" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434306 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0" exitCode=0 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434354 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401" exitCode=0 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434364 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3" exitCode=0 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434376 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d" exitCode=0 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434386 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381" exitCode=0 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434396 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3" exitCode=0 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434408 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab" exitCode=143 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434423 4788 generic.go:334] "Generic (PLEG): container finished" podID="baeca99f-afc8-4129-8110-51e79dd8b054" containerID="a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65" exitCode=143 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434506 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434550 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434572 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434585 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434600 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434626 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434646 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434659 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434668 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434675 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434683 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434691 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434698 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434705 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434713 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434721 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434731 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434743 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434752 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434759 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434769 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434778 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434787 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434794 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434804 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434850 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434859 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434756 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.434870 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435136 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435178 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435187 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435196 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435203 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435211 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435219 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435226 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435241 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435249 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435262 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cwgns" event={"ID":"baeca99f-afc8-4129-8110-51e79dd8b054","Type":"ContainerDied","Data":"c87fdf5d11a77f4ff6ffdb2328a38057d29f20709779bd6c141fd8dc7502e076"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435277 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435288 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435324 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435333 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435340 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435348 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435356 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435363 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435370 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.435377 4788 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.437173 4788 generic.go:334] "Generic (PLEG): container finished" podID="5e62642c-2473-488a-acf5-5f6903918e4f" containerID="43c0623a33e31f218d261ec1378059e54fbb12855d0d92c991df187179ebf1b1" exitCode=0 Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.437220 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerDied","Data":"43c0623a33e31f218d261ec1378059e54fbb12855d0d92c991df187179ebf1b1"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.437278 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"65a5820e72c3c58e4fe0ea106400e3fc62a8b87769379b710a83271c96bb8690"} Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.463293 4788 scope.go:117] "RemoveContainer" containerID="e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.490672 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cwgns"] Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.496260 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.497944 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cwgns"] Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.569697 4788 scope.go:117] "RemoveContainer" containerID="574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.594754 4788 scope.go:117] "RemoveContainer" containerID="498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.621589 4788 scope.go:117] "RemoveContainer" containerID="4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.641323 4788 scope.go:117] "RemoveContainer" containerID="a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.661522 4788 scope.go:117] "RemoveContainer" containerID="0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.691481 4788 scope.go:117] "RemoveContainer" containerID="41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.709052 4788 scope.go:117] "RemoveContainer" containerID="a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.739739 4788 scope.go:117] "RemoveContainer" containerID="68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.766551 4788 scope.go:117] "RemoveContainer" containerID="e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.767104 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": container with ID starting with e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0 not found: ID does not exist" containerID="e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.767237 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} err="failed to get container status \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": rpc error: code = NotFound desc = could not find container \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": container with ID starting with e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.767277 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.767606 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": container with ID starting with e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58 not found: ID does not exist" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.767639 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} err="failed to get container status \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": rpc error: code = NotFound desc = could not find container \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": container with ID starting with e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.767663 4788 scope.go:117] "RemoveContainer" containerID="574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.768176 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": container with ID starting with 574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401 not found: ID does not exist" containerID="574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.768233 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} err="failed to get container status \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": rpc error: code = NotFound desc = could not find container \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": container with ID starting with 574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.768283 4788 scope.go:117] "RemoveContainer" containerID="498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.768649 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": container with ID starting with 498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3 not found: ID does not exist" containerID="498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.768676 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} err="failed to get container status \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": rpc error: code = NotFound desc = could not find container \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": container with ID starting with 498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.768733 4788 scope.go:117] "RemoveContainer" containerID="4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.769199 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": container with ID starting with 4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d not found: ID does not exist" containerID="4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.769246 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} err="failed to get container status \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": rpc error: code = NotFound desc = could not find container \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": container with ID starting with 4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.769277 4788 scope.go:117] "RemoveContainer" containerID="a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.769628 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": container with ID starting with a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381 not found: ID does not exist" containerID="a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.769657 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} err="failed to get container status \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": rpc error: code = NotFound desc = could not find container \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": container with ID starting with a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.769673 4788 scope.go:117] "RemoveContainer" containerID="0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.770072 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": container with ID starting with 0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3 not found: ID does not exist" containerID="0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.770101 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} err="failed to get container status \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": rpc error: code = NotFound desc = could not find container \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": container with ID starting with 0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.770120 4788 scope.go:117] "RemoveContainer" containerID="41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.770477 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": container with ID starting with 41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab not found: ID does not exist" containerID="41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.770505 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} err="failed to get container status \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": rpc error: code = NotFound desc = could not find container \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": container with ID starting with 41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.770523 4788 scope.go:117] "RemoveContainer" containerID="a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.771066 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": container with ID starting with a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65 not found: ID does not exist" containerID="a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.771202 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} err="failed to get container status \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": rpc error: code = NotFound desc = could not find container \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": container with ID starting with a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.771224 4788 scope.go:117] "RemoveContainer" containerID="68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717" Oct 10 14:57:40 crc kubenswrapper[4788]: E1010 14:57:40.771526 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": container with ID starting with 68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717 not found: ID does not exist" containerID="68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.771557 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} err="failed to get container status \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": rpc error: code = NotFound desc = could not find container \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": container with ID starting with 68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.771579 4788 scope.go:117] "RemoveContainer" containerID="e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.772003 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} err="failed to get container status \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": rpc error: code = NotFound desc = could not find container \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": container with ID starting with e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.772028 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.772449 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} err="failed to get container status \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": rpc error: code = NotFound desc = could not find container \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": container with ID starting with e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.772479 4788 scope.go:117] "RemoveContainer" containerID="574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.772808 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} err="failed to get container status \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": rpc error: code = NotFound desc = could not find container \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": container with ID starting with 574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.772827 4788 scope.go:117] "RemoveContainer" containerID="498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.773156 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} err="failed to get container status \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": rpc error: code = NotFound desc = could not find container \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": container with ID starting with 498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.773188 4788 scope.go:117] "RemoveContainer" containerID="4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.773446 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} err="failed to get container status \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": rpc error: code = NotFound desc = could not find container \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": container with ID starting with 4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.773468 4788 scope.go:117] "RemoveContainer" containerID="a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.773822 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} err="failed to get container status \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": rpc error: code = NotFound desc = could not find container \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": container with ID starting with a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.773899 4788 scope.go:117] "RemoveContainer" containerID="0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.774230 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} err="failed to get container status \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": rpc error: code = NotFound desc = could not find container \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": container with ID starting with 0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.774254 4788 scope.go:117] "RemoveContainer" containerID="41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.774684 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} err="failed to get container status \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": rpc error: code = NotFound desc = could not find container \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": container with ID starting with 41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.774721 4788 scope.go:117] "RemoveContainer" containerID="a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.775433 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} err="failed to get container status \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": rpc error: code = NotFound desc = could not find container \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": container with ID starting with a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.775457 4788 scope.go:117] "RemoveContainer" containerID="68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.776005 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} err="failed to get container status \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": rpc error: code = NotFound desc = could not find container \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": container with ID starting with 68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.776036 4788 scope.go:117] "RemoveContainer" containerID="e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.777284 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} err="failed to get container status \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": rpc error: code = NotFound desc = could not find container \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": container with ID starting with e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.777311 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.777542 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} err="failed to get container status \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": rpc error: code = NotFound desc = could not find container \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": container with ID starting with e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.777561 4788 scope.go:117] "RemoveContainer" containerID="574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.777812 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} err="failed to get container status \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": rpc error: code = NotFound desc = could not find container \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": container with ID starting with 574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.777834 4788 scope.go:117] "RemoveContainer" containerID="498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.778268 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} err="failed to get container status \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": rpc error: code = NotFound desc = could not find container \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": container with ID starting with 498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.778291 4788 scope.go:117] "RemoveContainer" containerID="4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.778630 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} err="failed to get container status \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": rpc error: code = NotFound desc = could not find container \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": container with ID starting with 4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.778666 4788 scope.go:117] "RemoveContainer" containerID="a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.779020 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} err="failed to get container status \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": rpc error: code = NotFound desc = could not find container \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": container with ID starting with a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.779088 4788 scope.go:117] "RemoveContainer" containerID="0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.779454 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} err="failed to get container status \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": rpc error: code = NotFound desc = could not find container \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": container with ID starting with 0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.779478 4788 scope.go:117] "RemoveContainer" containerID="41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.779814 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} err="failed to get container status \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": rpc error: code = NotFound desc = could not find container \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": container with ID starting with 41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.779844 4788 scope.go:117] "RemoveContainer" containerID="a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.780190 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} err="failed to get container status \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": rpc error: code = NotFound desc = could not find container \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": container with ID starting with a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.780210 4788 scope.go:117] "RemoveContainer" containerID="68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.780588 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} err="failed to get container status \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": rpc error: code = NotFound desc = could not find container \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": container with ID starting with 68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.780623 4788 scope.go:117] "RemoveContainer" containerID="e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.780936 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0"} err="failed to get container status \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": rpc error: code = NotFound desc = could not find container \"e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0\": container with ID starting with e13dee60093349a6afd303f7bea8277a63a2f15668bc34a41bac27491f95a4b0 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.780962 4788 scope.go:117] "RemoveContainer" containerID="e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.781291 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58"} err="failed to get container status \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": rpc error: code = NotFound desc = could not find container \"e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58\": container with ID starting with e1300298295d684d64eb29c5fa891a087f0a235af08b1ba41871f3b65ee8be58 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.781316 4788 scope.go:117] "RemoveContainer" containerID="574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.781670 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401"} err="failed to get container status \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": rpc error: code = NotFound desc = could not find container \"574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401\": container with ID starting with 574b8f1a71dcfd1a87200c23eebcd49ff2615c6323518a289a00c21effd52401 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.781698 4788 scope.go:117] "RemoveContainer" containerID="498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.781923 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3"} err="failed to get container status \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": rpc error: code = NotFound desc = could not find container \"498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3\": container with ID starting with 498d1bbf3ab28010e73304f2636ecdf4a814b76e67501762f7ce50edfa8b46f3 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.781942 4788 scope.go:117] "RemoveContainer" containerID="4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.782286 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d"} err="failed to get container status \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": rpc error: code = NotFound desc = could not find container \"4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d\": container with ID starting with 4bfae1c12df57c5ff5405561959ae9b134b0c1cd54dc031a5ea8c1234357150d not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.782310 4788 scope.go:117] "RemoveContainer" containerID="a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.782585 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381"} err="failed to get container status \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": rpc error: code = NotFound desc = could not find container \"a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381\": container with ID starting with a39576297f3d440f2e6d7912b5be59f4d368efa90ee680fe2be228f4745ff381 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.782617 4788 scope.go:117] "RemoveContainer" containerID="0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.782881 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3"} err="failed to get container status \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": rpc error: code = NotFound desc = could not find container \"0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3\": container with ID starting with 0a29c35cfc77bfc0fac72d4981289af66f09773a2868d2eedc20bf0e875dc5c3 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.782904 4788 scope.go:117] "RemoveContainer" containerID="41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.783253 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab"} err="failed to get container status \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": rpc error: code = NotFound desc = could not find container \"41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab\": container with ID starting with 41307656f863c2d5f3b723214dd1b3856a3aea38640c2c7511fe5de8ec72e7ab not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.783278 4788 scope.go:117] "RemoveContainer" containerID="a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.783574 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65"} err="failed to get container status \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": rpc error: code = NotFound desc = could not find container \"a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65\": container with ID starting with a8e93fffa5d4c22d108e771c889235e27ef3e251ad1871aab6f68fc83ca23c65 not found: ID does not exist" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.783604 4788 scope.go:117] "RemoveContainer" containerID="68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717" Oct 10 14:57:40 crc kubenswrapper[4788]: I1010 14:57:40.783888 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717"} err="failed to get container status \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": rpc error: code = NotFound desc = could not find container \"68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717\": container with ID starting with 68c1708ce8fe0d7421a26ccebea14565a9c33f72d1bc22797fc52512b561c717 not found: ID does not exist" Oct 10 14:57:41 crc kubenswrapper[4788]: I1010 14:57:41.459068 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rzgfk_39cfa1ec-4912-4550-9ead-7f6113db2221/kube-multus/2.log" Oct 10 14:57:41 crc kubenswrapper[4788]: I1010 14:57:41.459783 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rzgfk" event={"ID":"39cfa1ec-4912-4550-9ead-7f6113db2221","Type":"ContainerStarted","Data":"4cf404e137d4fa44629161282a184d30e0384b6755e42424e344347b3efb90f9"} Oct 10 14:57:41 crc kubenswrapper[4788]: I1010 14:57:41.467325 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"e19f1104902e389a9b550e96bb9c4447dc39b14290cb2162e4e1bfe755ed9a78"} Oct 10 14:57:41 crc kubenswrapper[4788]: I1010 14:57:41.467380 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"27f38137f83e1482824d548a102e6807d70942955d8d0ecd09f57c452f41d686"} Oct 10 14:57:41 crc kubenswrapper[4788]: I1010 14:57:41.467394 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"baf94427d93be3726071f9dff52d1f77942ba5f83d2ffb0320dcb78ad6ff2b32"} Oct 10 14:57:41 crc kubenswrapper[4788]: I1010 14:57:41.467406 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"ebf765771b35d36080843104a8b5e32673235c55e26f3903c4e547ae0c243146"} Oct 10 14:57:41 crc kubenswrapper[4788]: I1010 14:57:41.467420 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"17675f46efe8e6dbbd2a98e4745d190b6468ee87b4b06e54ad4b7b340e95356d"} Oct 10 14:57:41 crc kubenswrapper[4788]: I1010 14:57:41.467435 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"117413cc895192ce0ced9d2fb2bff208415b5ed5cd34f41f1100aa9f6f9b08e1"} Oct 10 14:57:42 crc kubenswrapper[4788]: I1010 14:57:42.249317 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baeca99f-afc8-4129-8110-51e79dd8b054" path="/var/lib/kubelet/pods/baeca99f-afc8-4129-8110-51e79dd8b054/volumes" Oct 10 14:57:44 crc kubenswrapper[4788]: I1010 14:57:44.491290 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"544fd30c123dc051e55490887978e518bc5302b8bd29968a2ca03d8377c689de"} Oct 10 14:57:46 crc kubenswrapper[4788]: I1010 14:57:46.510339 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" event={"ID":"5e62642c-2473-488a-acf5-5f6903918e4f","Type":"ContainerStarted","Data":"e435be95ebca79b15fa8ad4289bfb25633f9121322de0342b9c3e667a7ad207b"} Oct 10 14:57:46 crc kubenswrapper[4788]: I1010 14:57:46.511429 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:46 crc kubenswrapper[4788]: I1010 14:57:46.537746 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" podStartSLOduration=7.537723152 podStartE2EDuration="7.537723152s" podCreationTimestamp="2025-10-10 14:57:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:57:46.533525174 +0000 UTC m=+768.983240742" watchObservedRunningTime="2025-10-10 14:57:46.537723152 +0000 UTC m=+768.987438700" Oct 10 14:57:46 crc kubenswrapper[4788]: I1010 14:57:46.542593 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:47 crc kubenswrapper[4788]: I1010 14:57:47.519012 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:47 crc kubenswrapper[4788]: I1010 14:57:47.519066 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:47 crc kubenswrapper[4788]: I1010 14:57:47.549619 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:57:48 crc kubenswrapper[4788]: I1010 14:57:48.945635 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-8442z"] Oct 10 14:57:48 crc kubenswrapper[4788]: I1010 14:57:48.947710 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:48 crc kubenswrapper[4788]: I1010 14:57:48.951830 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 10 14:57:48 crc kubenswrapper[4788]: I1010 14:57:48.951957 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 10 14:57:48 crc kubenswrapper[4788]: I1010 14:57:48.954690 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 10 14:57:48 crc kubenswrapper[4788]: I1010 14:57:48.954840 4788 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-psrrw" Oct 10 14:57:48 crc kubenswrapper[4788]: I1010 14:57:48.955971 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-8442z"] Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.137870 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1d08229c-455b-4b6a-a904-f114306b9e1b-node-mnt\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.137948 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1d08229c-455b-4b6a-a904-f114306b9e1b-crc-storage\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.138056 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq5f9\" (UniqueName: \"kubernetes.io/projected/1d08229c-455b-4b6a-a904-f114306b9e1b-kube-api-access-tq5f9\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.239830 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1d08229c-455b-4b6a-a904-f114306b9e1b-node-mnt\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.239909 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1d08229c-455b-4b6a-a904-f114306b9e1b-crc-storage\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.240019 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq5f9\" (UniqueName: \"kubernetes.io/projected/1d08229c-455b-4b6a-a904-f114306b9e1b-kube-api-access-tq5f9\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.240455 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1d08229c-455b-4b6a-a904-f114306b9e1b-node-mnt\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.241372 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1d08229c-455b-4b6a-a904-f114306b9e1b-crc-storage\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.280822 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq5f9\" (UniqueName: \"kubernetes.io/projected/1d08229c-455b-4b6a-a904-f114306b9e1b-kube-api-access-tq5f9\") pod \"crc-storage-crc-8442z\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.281419 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: E1010 14:57:49.320969 4788 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-8442z_crc-storage_1d08229c-455b-4b6a-a904-f114306b9e1b_0(811ec0330fd93604745efec195953a052f6136705c8e448e3195e75ba85278fb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 10 14:57:49 crc kubenswrapper[4788]: E1010 14:57:49.321103 4788 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-8442z_crc-storage_1d08229c-455b-4b6a-a904-f114306b9e1b_0(811ec0330fd93604745efec195953a052f6136705c8e448e3195e75ba85278fb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: E1010 14:57:49.321161 4788 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-8442z_crc-storage_1d08229c-455b-4b6a-a904-f114306b9e1b_0(811ec0330fd93604745efec195953a052f6136705c8e448e3195e75ba85278fb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: E1010 14:57:49.321231 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-8442z_crc-storage(1d08229c-455b-4b6a-a904-f114306b9e1b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-8442z_crc-storage(1d08229c-455b-4b6a-a904-f114306b9e1b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-8442z_crc-storage_1d08229c-455b-4b6a-a904-f114306b9e1b_0(811ec0330fd93604745efec195953a052f6136705c8e448e3195e75ba85278fb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-8442z" podUID="1d08229c-455b-4b6a-a904-f114306b9e1b" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.532796 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: I1010 14:57:49.534053 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: E1010 14:57:49.567212 4788 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-8442z_crc-storage_1d08229c-455b-4b6a-a904-f114306b9e1b_0(eb76f8fa13877a65330945462cb6947193a0c92042d61565b28d891fbfe19da8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 10 14:57:49 crc kubenswrapper[4788]: E1010 14:57:49.567299 4788 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-8442z_crc-storage_1d08229c-455b-4b6a-a904-f114306b9e1b_0(eb76f8fa13877a65330945462cb6947193a0c92042d61565b28d891fbfe19da8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: E1010 14:57:49.567331 4788 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-8442z_crc-storage_1d08229c-455b-4b6a-a904-f114306b9e1b_0(eb76f8fa13877a65330945462cb6947193a0c92042d61565b28d891fbfe19da8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:57:49 crc kubenswrapper[4788]: E1010 14:57:49.567392 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-8442z_crc-storage(1d08229c-455b-4b6a-a904-f114306b9e1b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-8442z_crc-storage(1d08229c-455b-4b6a-a904-f114306b9e1b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-8442z_crc-storage_1d08229c-455b-4b6a-a904-f114306b9e1b_0(eb76f8fa13877a65330945462cb6947193a0c92042d61565b28d891fbfe19da8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-8442z" podUID="1d08229c-455b-4b6a-a904-f114306b9e1b" Oct 10 14:58:02 crc kubenswrapper[4788]: I1010 14:58:02.233980 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:58:02 crc kubenswrapper[4788]: I1010 14:58:02.235951 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:58:02 crc kubenswrapper[4788]: I1010 14:58:02.748232 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-8442z"] Oct 10 14:58:02 crc kubenswrapper[4788]: I1010 14:58:02.767601 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 14:58:03 crc kubenswrapper[4788]: I1010 14:58:03.630859 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8442z" event={"ID":"1d08229c-455b-4b6a-a904-f114306b9e1b","Type":"ContainerStarted","Data":"71147382cd437a11ab8217336f8c83d3eb3d7016e5eb09ea48d9c6a1a9783191"} Oct 10 14:58:04 crc kubenswrapper[4788]: I1010 14:58:04.639590 4788 generic.go:334] "Generic (PLEG): container finished" podID="1d08229c-455b-4b6a-a904-f114306b9e1b" containerID="15278051e22f4349d19272666fe9aef820f9cbc1325052282f79fa116cecb9a4" exitCode=0 Oct 10 14:58:04 crc kubenswrapper[4788]: I1010 14:58:04.639660 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8442z" event={"ID":"1d08229c-455b-4b6a-a904-f114306b9e1b","Type":"ContainerDied","Data":"15278051e22f4349d19272666fe9aef820f9cbc1325052282f79fa116cecb9a4"} Oct 10 14:58:05 crc kubenswrapper[4788]: I1010 14:58:05.996931 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.194576 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1d08229c-455b-4b6a-a904-f114306b9e1b-crc-storage\") pod \"1d08229c-455b-4b6a-a904-f114306b9e1b\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.195357 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq5f9\" (UniqueName: \"kubernetes.io/projected/1d08229c-455b-4b6a-a904-f114306b9e1b-kube-api-access-tq5f9\") pod \"1d08229c-455b-4b6a-a904-f114306b9e1b\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.195543 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1d08229c-455b-4b6a-a904-f114306b9e1b-node-mnt\") pod \"1d08229c-455b-4b6a-a904-f114306b9e1b\" (UID: \"1d08229c-455b-4b6a-a904-f114306b9e1b\") " Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.195772 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1d08229c-455b-4b6a-a904-f114306b9e1b-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "1d08229c-455b-4b6a-a904-f114306b9e1b" (UID: "1d08229c-455b-4b6a-a904-f114306b9e1b"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.196193 4788 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1d08229c-455b-4b6a-a904-f114306b9e1b-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.203467 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d08229c-455b-4b6a-a904-f114306b9e1b-kube-api-access-tq5f9" (OuterVolumeSpecName: "kube-api-access-tq5f9") pod "1d08229c-455b-4b6a-a904-f114306b9e1b" (UID: "1d08229c-455b-4b6a-a904-f114306b9e1b"). InnerVolumeSpecName "kube-api-access-tq5f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.223490 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d08229c-455b-4b6a-a904-f114306b9e1b-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "1d08229c-455b-4b6a-a904-f114306b9e1b" (UID: "1d08229c-455b-4b6a-a904-f114306b9e1b"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.298272 4788 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1d08229c-455b-4b6a-a904-f114306b9e1b-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.298350 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq5f9\" (UniqueName: \"kubernetes.io/projected/1d08229c-455b-4b6a-a904-f114306b9e1b-kube-api-access-tq5f9\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.657498 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8442z" event={"ID":"1d08229c-455b-4b6a-a904-f114306b9e1b","Type":"ContainerDied","Data":"71147382cd437a11ab8217336f8c83d3eb3d7016e5eb09ea48d9c6a1a9783191"} Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.657580 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8442z" Oct 10 14:58:06 crc kubenswrapper[4788]: I1010 14:58:06.657591 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71147382cd437a11ab8217336f8c83d3eb3d7016e5eb09ea48d9c6a1a9783191" Oct 10 14:58:10 crc kubenswrapper[4788]: I1010 14:58:10.180356 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8v4sx" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.795913 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd"] Oct 10 14:58:14 crc kubenswrapper[4788]: E1010 14:58:14.797556 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d08229c-455b-4b6a-a904-f114306b9e1b" containerName="storage" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.797577 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d08229c-455b-4b6a-a904-f114306b9e1b" containerName="storage" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.797697 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d08229c-455b-4b6a-a904-f114306b9e1b" containerName="storage" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.798509 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.800618 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.813227 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd"] Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.831557 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.831639 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xznsw\" (UniqueName: \"kubernetes.io/projected/33653f49-1164-45f1-90b9-d22e4c531b5e-kube-api-access-xznsw\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.831776 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.932995 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.933088 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xznsw\" (UniqueName: \"kubernetes.io/projected/33653f49-1164-45f1-90b9-d22e4c531b5e-kube-api-access-xznsw\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.933174 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.933512 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.933531 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:14 crc kubenswrapper[4788]: I1010 14:58:14.960098 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xznsw\" (UniqueName: \"kubernetes.io/projected/33653f49-1164-45f1-90b9-d22e4c531b5e-kube-api-access-xznsw\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:15 crc kubenswrapper[4788]: I1010 14:58:15.120658 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:15 crc kubenswrapper[4788]: I1010 14:58:15.325004 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd"] Oct 10 14:58:15 crc kubenswrapper[4788]: I1010 14:58:15.722265 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" event={"ID":"33653f49-1164-45f1-90b9-d22e4c531b5e","Type":"ContainerStarted","Data":"8867d7efae6b8db985e409717dc8794beb564b6073a95790c5d47ce714c0174d"} Oct 10 14:58:15 crc kubenswrapper[4788]: I1010 14:58:15.722824 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" event={"ID":"33653f49-1164-45f1-90b9-d22e4c531b5e","Type":"ContainerStarted","Data":"976c931346df6338e152fa930533e67f7637589c69be4304c6a9e183f7a05921"} Oct 10 14:58:16 crc kubenswrapper[4788]: I1010 14:58:16.731513 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" event={"ID":"33653f49-1164-45f1-90b9-d22e4c531b5e","Type":"ContainerDied","Data":"8867d7efae6b8db985e409717dc8794beb564b6073a95790c5d47ce714c0174d"} Oct 10 14:58:16 crc kubenswrapper[4788]: I1010 14:58:16.731430 4788 generic.go:334] "Generic (PLEG): container finished" podID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerID="8867d7efae6b8db985e409717dc8794beb564b6073a95790c5d47ce714c0174d" exitCode=0 Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.035239 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x9zrb"] Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.037581 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.047837 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9zrb"] Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.162788 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-catalog-content\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.163693 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-utilities\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.163743 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8vw6\" (UniqueName: \"kubernetes.io/projected/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-kube-api-access-r8vw6\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.264589 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8vw6\" (UniqueName: \"kubernetes.io/projected/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-kube-api-access-r8vw6\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.264692 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-utilities\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.264781 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-catalog-content\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.265494 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-catalog-content\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.265797 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-utilities\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.296672 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8vw6\" (UniqueName: \"kubernetes.io/projected/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-kube-api-access-r8vw6\") pod \"redhat-operators-x9zrb\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.374956 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:17 crc kubenswrapper[4788]: I1010 14:58:17.802645 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9zrb"] Oct 10 14:58:17 crc kubenswrapper[4788]: W1010 14:58:17.812798 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf9167a4_bbaf_4e8a_bc27_ad7d6ee862a0.slice/crio-b86e7abca27081bb3cb25fbbb07ce648706e74462d36fd49c51866505482cbc8 WatchSource:0}: Error finding container b86e7abca27081bb3cb25fbbb07ce648706e74462d36fd49c51866505482cbc8: Status 404 returned error can't find the container with id b86e7abca27081bb3cb25fbbb07ce648706e74462d36fd49c51866505482cbc8 Oct 10 14:58:18 crc kubenswrapper[4788]: I1010 14:58:18.744959 4788 generic.go:334] "Generic (PLEG): container finished" podID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerID="63a9c53a08bb60bbec85c4f38a751851f01f381702e1fbc4f227a65e0f91890a" exitCode=0 Oct 10 14:58:18 crc kubenswrapper[4788]: I1010 14:58:18.745059 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" event={"ID":"33653f49-1164-45f1-90b9-d22e4c531b5e","Type":"ContainerDied","Data":"63a9c53a08bb60bbec85c4f38a751851f01f381702e1fbc4f227a65e0f91890a"} Oct 10 14:58:18 crc kubenswrapper[4788]: I1010 14:58:18.747059 4788 generic.go:334] "Generic (PLEG): container finished" podID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerID="380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df" exitCode=0 Oct 10 14:58:18 crc kubenswrapper[4788]: I1010 14:58:18.747117 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9zrb" event={"ID":"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0","Type":"ContainerDied","Data":"380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df"} Oct 10 14:58:18 crc kubenswrapper[4788]: I1010 14:58:18.747201 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9zrb" event={"ID":"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0","Type":"ContainerStarted","Data":"b86e7abca27081bb3cb25fbbb07ce648706e74462d36fd49c51866505482cbc8"} Oct 10 14:58:19 crc kubenswrapper[4788]: I1010 14:58:19.764098 4788 generic.go:334] "Generic (PLEG): container finished" podID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerID="505229b95b3dbbe548b5a326e9900a9ecb557f6c32711f6b7429db1684ebd6eb" exitCode=0 Oct 10 14:58:19 crc kubenswrapper[4788]: I1010 14:58:19.764225 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" event={"ID":"33653f49-1164-45f1-90b9-d22e4c531b5e","Type":"ContainerDied","Data":"505229b95b3dbbe548b5a326e9900a9ecb557f6c32711f6b7429db1684ebd6eb"} Oct 10 14:58:20 crc kubenswrapper[4788]: I1010 14:58:20.778268 4788 generic.go:334] "Generic (PLEG): container finished" podID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerID="f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261" exitCode=0 Oct 10 14:58:20 crc kubenswrapper[4788]: I1010 14:58:20.778389 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9zrb" event={"ID":"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0","Type":"ContainerDied","Data":"f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261"} Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.078227 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.134428 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xznsw\" (UniqueName: \"kubernetes.io/projected/33653f49-1164-45f1-90b9-d22e4c531b5e-kube-api-access-xznsw\") pod \"33653f49-1164-45f1-90b9-d22e4c531b5e\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.134474 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-util\") pod \"33653f49-1164-45f1-90b9-d22e4c531b5e\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.134521 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-bundle\") pod \"33653f49-1164-45f1-90b9-d22e4c531b5e\" (UID: \"33653f49-1164-45f1-90b9-d22e4c531b5e\") " Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.135072 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-bundle" (OuterVolumeSpecName: "bundle") pod "33653f49-1164-45f1-90b9-d22e4c531b5e" (UID: "33653f49-1164-45f1-90b9-d22e4c531b5e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.139738 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33653f49-1164-45f1-90b9-d22e4c531b5e-kube-api-access-xznsw" (OuterVolumeSpecName: "kube-api-access-xznsw") pod "33653f49-1164-45f1-90b9-d22e4c531b5e" (UID: "33653f49-1164-45f1-90b9-d22e4c531b5e"). InnerVolumeSpecName "kube-api-access-xznsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.180928 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-util" (OuterVolumeSpecName: "util") pod "33653f49-1164-45f1-90b9-d22e4c531b5e" (UID: "33653f49-1164-45f1-90b9-d22e4c531b5e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.236168 4788 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.236197 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xznsw\" (UniqueName: \"kubernetes.io/projected/33653f49-1164-45f1-90b9-d22e4c531b5e-kube-api-access-xznsw\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.236210 4788 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33653f49-1164-45f1-90b9-d22e4c531b5e-util\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.785153 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" event={"ID":"33653f49-1164-45f1-90b9-d22e4c531b5e","Type":"ContainerDied","Data":"976c931346df6338e152fa930533e67f7637589c69be4304c6a9e183f7a05921"} Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.785439 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="976c931346df6338e152fa930533e67f7637589c69be4304c6a9e183f7a05921" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.785206 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd" Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.787931 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9zrb" event={"ID":"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0","Type":"ContainerStarted","Data":"0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce"} Oct 10 14:58:21 crc kubenswrapper[4788]: I1010 14:58:21.807312 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x9zrb" podStartSLOduration=2.32755699 podStartE2EDuration="4.807289449s" podCreationTimestamp="2025-10-10 14:58:17 +0000 UTC" firstStartedPulling="2025-10-10 14:58:18.749369173 +0000 UTC m=+801.199084731" lastFinishedPulling="2025-10-10 14:58:21.229101632 +0000 UTC m=+803.678817190" observedRunningTime="2025-10-10 14:58:21.804218083 +0000 UTC m=+804.253933631" watchObservedRunningTime="2025-10-10 14:58:21.807289449 +0000 UTC m=+804.257004997" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.031685 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mfhz2"] Oct 10 14:58:24 crc kubenswrapper[4788]: E1010 14:58:24.032001 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerName="pull" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.032019 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerName="pull" Oct 10 14:58:24 crc kubenswrapper[4788]: E1010 14:58:24.032060 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerName="extract" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.032070 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerName="extract" Oct 10 14:58:24 crc kubenswrapper[4788]: E1010 14:58:24.032087 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerName="util" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.032096 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerName="util" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.032227 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="33653f49-1164-45f1-90b9-d22e4c531b5e" containerName="extract" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.032943 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.051809 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mfhz2"] Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.069845 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-utilities\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.069915 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-catalog-content\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.069998 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc8hc\" (UniqueName: \"kubernetes.io/projected/15beebd9-6640-4117-ab2a-047a42b83a79-kube-api-access-cc8hc\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.171072 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-utilities\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.171120 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-catalog-content\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.171201 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc8hc\" (UniqueName: \"kubernetes.io/projected/15beebd9-6640-4117-ab2a-047a42b83a79-kube-api-access-cc8hc\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.171610 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-utilities\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.171764 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-catalog-content\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.197855 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc8hc\" (UniqueName: \"kubernetes.io/projected/15beebd9-6640-4117-ab2a-047a42b83a79-kube-api-access-cc8hc\") pod \"community-operators-mfhz2\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.365319 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:24 crc kubenswrapper[4788]: I1010 14:58:24.853154 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mfhz2"] Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.358979 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-k6czn"] Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.360509 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-k6czn" Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.362840 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.364186 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.366637 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bbmqn" Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.373111 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-k6czn"] Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.392286 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg7t7\" (UniqueName: \"kubernetes.io/projected/a6752f9a-c8ce-4c2b-b234-b5229f430a5e-kube-api-access-rg7t7\") pod \"nmstate-operator-858ddd8f98-k6czn\" (UID: \"a6752f9a-c8ce-4c2b-b234-b5229f430a5e\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-k6czn" Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.493774 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg7t7\" (UniqueName: \"kubernetes.io/projected/a6752f9a-c8ce-4c2b-b234-b5229f430a5e-kube-api-access-rg7t7\") pod \"nmstate-operator-858ddd8f98-k6czn\" (UID: \"a6752f9a-c8ce-4c2b-b234-b5229f430a5e\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-k6czn" Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.513215 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg7t7\" (UniqueName: \"kubernetes.io/projected/a6752f9a-c8ce-4c2b-b234-b5229f430a5e-kube-api-access-rg7t7\") pod \"nmstate-operator-858ddd8f98-k6czn\" (UID: \"a6752f9a-c8ce-4c2b-b234-b5229f430a5e\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-k6czn" Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.677738 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-k6czn" Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.811714 4788 generic.go:334] "Generic (PLEG): container finished" podID="15beebd9-6640-4117-ab2a-047a42b83a79" containerID="06ef843888d1cd292ed6c13eef398242e61ee630baf38acfa7fb7ac7aa461c92" exitCode=0 Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.812119 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfhz2" event={"ID":"15beebd9-6640-4117-ab2a-047a42b83a79","Type":"ContainerDied","Data":"06ef843888d1cd292ed6c13eef398242e61ee630baf38acfa7fb7ac7aa461c92"} Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.812166 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfhz2" event={"ID":"15beebd9-6640-4117-ab2a-047a42b83a79","Type":"ContainerStarted","Data":"396ea2658e1ec1961fb6a12244ab83b8afaa081baa92c5934c0380e650fa4c61"} Oct 10 14:58:25 crc kubenswrapper[4788]: I1010 14:58:25.867200 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-k6czn"] Oct 10 14:58:25 crc kubenswrapper[4788]: W1010 14:58:25.877577 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6752f9a_c8ce_4c2b_b234_b5229f430a5e.slice/crio-379a722bf311e7fc8030248ec10ec8998195b2d5d28fd3eee702b6f6903f0b15 WatchSource:0}: Error finding container 379a722bf311e7fc8030248ec10ec8998195b2d5d28fd3eee702b6f6903f0b15: Status 404 returned error can't find the container with id 379a722bf311e7fc8030248ec10ec8998195b2d5d28fd3eee702b6f6903f0b15 Oct 10 14:58:26 crc kubenswrapper[4788]: I1010 14:58:26.818503 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-k6czn" event={"ID":"a6752f9a-c8ce-4c2b-b234-b5229f430a5e","Type":"ContainerStarted","Data":"379a722bf311e7fc8030248ec10ec8998195b2d5d28fd3eee702b6f6903f0b15"} Oct 10 14:58:27 crc kubenswrapper[4788]: I1010 14:58:27.375942 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:27 crc kubenswrapper[4788]: I1010 14:58:27.376605 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:27 crc kubenswrapper[4788]: I1010 14:58:27.456607 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:27 crc kubenswrapper[4788]: I1010 14:58:27.828596 4788 generic.go:334] "Generic (PLEG): container finished" podID="15beebd9-6640-4117-ab2a-047a42b83a79" containerID="60016f0d06bc1b87e425bc41d54071788445976c28e369c57c3206db42e399ec" exitCode=0 Oct 10 14:58:27 crc kubenswrapper[4788]: I1010 14:58:27.828653 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfhz2" event={"ID":"15beebd9-6640-4117-ab2a-047a42b83a79","Type":"ContainerDied","Data":"60016f0d06bc1b87e425bc41d54071788445976c28e369c57c3206db42e399ec"} Oct 10 14:58:27 crc kubenswrapper[4788]: I1010 14:58:27.871005 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:28 crc kubenswrapper[4788]: I1010 14:58:28.835021 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfhz2" event={"ID":"15beebd9-6640-4117-ab2a-047a42b83a79","Type":"ContainerStarted","Data":"1b68642f9c8585b60b7f3c1afd9c5cba67cbfccdc8d3be7e5de8c9bc8555251e"} Oct 10 14:58:28 crc kubenswrapper[4788]: I1010 14:58:28.839173 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-k6czn" event={"ID":"a6752f9a-c8ce-4c2b-b234-b5229f430a5e","Type":"ContainerStarted","Data":"5359c77e4bb78fcdd0bb0dddc0358ef027c084e23d98a8a616795dff20909818"} Oct 10 14:58:28 crc kubenswrapper[4788]: I1010 14:58:28.855791 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mfhz2" podStartSLOduration=2.096732447 podStartE2EDuration="4.855773917s" podCreationTimestamp="2025-10-10 14:58:24 +0000 UTC" firstStartedPulling="2025-10-10 14:58:25.814182346 +0000 UTC m=+808.263897894" lastFinishedPulling="2025-10-10 14:58:28.573223816 +0000 UTC m=+811.022939364" observedRunningTime="2025-10-10 14:58:28.853335068 +0000 UTC m=+811.303050606" watchObservedRunningTime="2025-10-10 14:58:28.855773917 +0000 UTC m=+811.305489465" Oct 10 14:58:28 crc kubenswrapper[4788]: I1010 14:58:28.873935 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-k6czn" podStartSLOduration=1.179645542 podStartE2EDuration="3.873904073s" podCreationTimestamp="2025-10-10 14:58:25 +0000 UTC" firstStartedPulling="2025-10-10 14:58:25.88128419 +0000 UTC m=+808.330999738" lastFinishedPulling="2025-10-10 14:58:28.575542711 +0000 UTC m=+811.025258269" observedRunningTime="2025-10-10 14:58:28.87310812 +0000 UTC m=+811.322823668" watchObservedRunningTime="2025-10-10 14:58:28.873904073 +0000 UTC m=+811.323619621" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.221254 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9zrb"] Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.221731 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x9zrb" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerName="registry-server" containerID="cri-o://0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce" gracePeriod=2 Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.709446 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.778871 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-catalog-content\") pod \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.779029 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-utilities\") pod \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.779055 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8vw6\" (UniqueName: \"kubernetes.io/projected/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-kube-api-access-r8vw6\") pod \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\" (UID: \"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0\") " Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.780068 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-utilities" (OuterVolumeSpecName: "utilities") pod "af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" (UID: "af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.784317 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-kube-api-access-r8vw6" (OuterVolumeSpecName: "kube-api-access-r8vw6") pod "af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" (UID: "af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0"). InnerVolumeSpecName "kube-api-access-r8vw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.858240 4788 generic.go:334] "Generic (PLEG): container finished" podID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerID="0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce" exitCode=0 Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.858285 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9zrb" event={"ID":"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0","Type":"ContainerDied","Data":"0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce"} Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.858315 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9zrb" event={"ID":"af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0","Type":"ContainerDied","Data":"b86e7abca27081bb3cb25fbbb07ce648706e74462d36fd49c51866505482cbc8"} Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.858333 4788 scope.go:117] "RemoveContainer" containerID="0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.858360 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9zrb" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.873907 4788 scope.go:117] "RemoveContainer" containerID="f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.880441 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.880472 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8vw6\" (UniqueName: \"kubernetes.io/projected/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-kube-api-access-r8vw6\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.890344 4788 scope.go:117] "RemoveContainer" containerID="380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.905477 4788 scope.go:117] "RemoveContainer" containerID="0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce" Oct 10 14:58:31 crc kubenswrapper[4788]: E1010 14:58:31.906130 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce\": container with ID starting with 0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce not found: ID does not exist" containerID="0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.906262 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce"} err="failed to get container status \"0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce\": rpc error: code = NotFound desc = could not find container \"0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce\": container with ID starting with 0c766524e475b272cd91c3c0b30f4f4d5ba9cc08ad0693ba71c3b2a29f77b2ce not found: ID does not exist" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.906341 4788 scope.go:117] "RemoveContainer" containerID="f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261" Oct 10 14:58:31 crc kubenswrapper[4788]: E1010 14:58:31.906994 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261\": container with ID starting with f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261 not found: ID does not exist" containerID="f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.907058 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261"} err="failed to get container status \"f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261\": rpc error: code = NotFound desc = could not find container \"f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261\": container with ID starting with f4b903c6f108f1bb5092560eafa83eeabab575fe0321d488a2a79eefe8c74261 not found: ID does not exist" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.907094 4788 scope.go:117] "RemoveContainer" containerID="380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df" Oct 10 14:58:31 crc kubenswrapper[4788]: E1010 14:58:31.907559 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df\": container with ID starting with 380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df not found: ID does not exist" containerID="380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.907684 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df"} err="failed to get container status \"380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df\": rpc error: code = NotFound desc = could not find container \"380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df\": container with ID starting with 380124ee9f98e4eb0fe6207d51ba163144b4a0da645fce3e1f3f91ccd8ffd9df not found: ID does not exist" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.912226 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" (UID: "af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:58:31 crc kubenswrapper[4788]: I1010 14:58:31.982104 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:32 crc kubenswrapper[4788]: I1010 14:58:32.186343 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9zrb"] Oct 10 14:58:32 crc kubenswrapper[4788]: I1010 14:58:32.192912 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x9zrb"] Oct 10 14:58:32 crc kubenswrapper[4788]: I1010 14:58:32.245086 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" path="/var/lib/kubelet/pods/af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0/volumes" Oct 10 14:58:34 crc kubenswrapper[4788]: I1010 14:58:34.366316 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:34 crc kubenswrapper[4788]: I1010 14:58:34.366428 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:34 crc kubenswrapper[4788]: I1010 14:58:34.437307 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:34 crc kubenswrapper[4788]: I1010 14:58:34.927650 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.509645 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md"] Oct 10 14:58:35 crc kubenswrapper[4788]: E1010 14:58:35.509997 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerName="registry-server" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.510022 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerName="registry-server" Oct 10 14:58:35 crc kubenswrapper[4788]: E1010 14:58:35.510042 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerName="extract-utilities" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.510051 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerName="extract-utilities" Oct 10 14:58:35 crc kubenswrapper[4788]: E1010 14:58:35.510070 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerName="extract-content" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.510082 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerName="extract-content" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.510245 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="af9167a4-bbaf-4e8a-bc27-ad7d6ee862a0" containerName="registry-server" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.510827 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.512828 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv"] Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.513944 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.515415 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.516266 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xk4jc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.536484 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv"] Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.539050 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-lg5x2"] Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.540118 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.541491 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md"] Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.542105 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2njhj\" (UniqueName: \"kubernetes.io/projected/c7329403-a7f9-45bb-93e9-7fdc91953d06-kube-api-access-2njhj\") pod \"nmstate-metrics-fdff9cb8d-7crrv\" (UID: \"c7329403-a7f9-45bb-93e9-7fdc91953d06\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.542199 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wd88\" (UniqueName: \"kubernetes.io/projected/d3b269c8-9fb0-4982-a512-eb5e91d37a97-kube-api-access-8wd88\") pod \"nmstate-webhook-6cdbc54649-5q4md\" (UID: \"d3b269c8-9fb0-4982-a512-eb5e91d37a97\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.542248 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d3b269c8-9fb0-4982-a512-eb5e91d37a97-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-5q4md\" (UID: \"d3b269c8-9fb0-4982-a512-eb5e91d37a97\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.644084 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-nmstate-lock\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.644331 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2njhj\" (UniqueName: \"kubernetes.io/projected/c7329403-a7f9-45bb-93e9-7fdc91953d06-kube-api-access-2njhj\") pod \"nmstate-metrics-fdff9cb8d-7crrv\" (UID: \"c7329403-a7f9-45bb-93e9-7fdc91953d06\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.644430 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-dbus-socket\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.644656 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wd88\" (UniqueName: \"kubernetes.io/projected/d3b269c8-9fb0-4982-a512-eb5e91d37a97-kube-api-access-8wd88\") pod \"nmstate-webhook-6cdbc54649-5q4md\" (UID: \"d3b269c8-9fb0-4982-a512-eb5e91d37a97\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.644721 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d3b269c8-9fb0-4982-a512-eb5e91d37a97-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-5q4md\" (UID: \"d3b269c8-9fb0-4982-a512-eb5e91d37a97\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.644797 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-ovs-socket\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.644815 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lzwf\" (UniqueName: \"kubernetes.io/projected/34af49c2-e879-487f-b92e-c178ea679f4d-kube-api-access-4lzwf\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.667874 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d3b269c8-9fb0-4982-a512-eb5e91d37a97-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-5q4md\" (UID: \"d3b269c8-9fb0-4982-a512-eb5e91d37a97\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.671985 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2njhj\" (UniqueName: \"kubernetes.io/projected/c7329403-a7f9-45bb-93e9-7fdc91953d06-kube-api-access-2njhj\") pod \"nmstate-metrics-fdff9cb8d-7crrv\" (UID: \"c7329403-a7f9-45bb-93e9-7fdc91953d06\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.679186 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wd88\" (UniqueName: \"kubernetes.io/projected/d3b269c8-9fb0-4982-a512-eb5e91d37a97-kube-api-access-8wd88\") pod \"nmstate-webhook-6cdbc54649-5q4md\" (UID: \"d3b269c8-9fb0-4982-a512-eb5e91d37a97\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.682381 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc"] Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.683213 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc"] Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.683319 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.685781 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-5k55g" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.685956 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.686082 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746329 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-nmstate-lock\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746405 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flcrn\" (UniqueName: \"kubernetes.io/projected/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-kube-api-access-flcrn\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746448 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-dbus-socket\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746473 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746524 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-ovs-socket\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746541 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lzwf\" (UniqueName: \"kubernetes.io/projected/34af49c2-e879-487f-b92e-c178ea679f4d-kube-api-access-4lzwf\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746567 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746666 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-nmstate-lock\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.746916 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-ovs-socket\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.747087 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/34af49c2-e879-487f-b92e-c178ea679f4d-dbus-socket\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.770917 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lzwf\" (UniqueName: \"kubernetes.io/projected/34af49c2-e879-487f-b92e-c178ea679f4d-kube-api-access-4lzwf\") pod \"nmstate-handler-lg5x2\" (UID: \"34af49c2-e879-487f-b92e-c178ea679f4d\") " pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.837271 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.846252 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-db8dd7775-q6mmf"] Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.847065 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.847402 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.847452 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flcrn\" (UniqueName: \"kubernetes.io/projected/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-kube-api-access-flcrn\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.847480 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.850390 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.850804 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.855114 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.862512 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.872220 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-db8dd7775-q6mmf"] Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.882262 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flcrn\" (UniqueName: \"kubernetes.io/projected/c5e5ab85-6606-4e44-bc7b-8fe2df836f49-kube-api-access-flcrn\") pod \"nmstate-console-plugin-6b874cbd85-d78sc\" (UID: \"c5e5ab85-6606-4e44-bc7b-8fe2df836f49\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.949733 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-oauth-config\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.949808 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-service-ca\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.949885 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-oauth-serving-cert\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.949938 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-config\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.949968 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t42n9\" (UniqueName: \"kubernetes.io/projected/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-kube-api-access-t42n9\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.950022 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-serving-cert\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:35 crc kubenswrapper[4788]: I1010 14:58:35.950093 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-trusted-ca-bundle\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.026211 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.051272 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-oauth-config\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.051316 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-service-ca\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.051350 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-oauth-serving-cert\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.051366 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-config\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.051382 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t42n9\" (UniqueName: \"kubernetes.io/projected/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-kube-api-access-t42n9\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.051404 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-serving-cert\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.051438 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-trusted-ca-bundle\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.052543 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-service-ca\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.052642 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-config\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.052938 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-trusted-ca-bundle\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.053081 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-oauth-serving-cert\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.060290 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-oauth-config\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.065972 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-console-serving-cert\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.073790 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t42n9\" (UniqueName: \"kubernetes.io/projected/ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87-kube-api-access-t42n9\") pod \"console-db8dd7775-q6mmf\" (UID: \"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87\") " pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.097630 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md"] Oct 10 14:58:36 crc kubenswrapper[4788]: W1010 14:58:36.099851 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3b269c8_9fb0_4982_a512_eb5e91d37a97.slice/crio-446b250283e587078768e4de9bdc972343f5d481511efa03e4feeffee4794a38 WatchSource:0}: Error finding container 446b250283e587078768e4de9bdc972343f5d481511efa03e4feeffee4794a38: Status 404 returned error can't find the container with id 446b250283e587078768e4de9bdc972343f5d481511efa03e4feeffee4794a38 Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.160508 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv"] Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.212676 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.396764 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-db8dd7775-q6mmf"] Oct 10 14:58:36 crc kubenswrapper[4788]: W1010 14:58:36.403088 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad9b84c3_2c65_4b1d_a1b7_96d3ef6bce87.slice/crio-334450f2c8760f6a70e4305009afa9861d4386fd53ac4de80a6609df962e6090 WatchSource:0}: Error finding container 334450f2c8760f6a70e4305009afa9861d4386fd53ac4de80a6609df962e6090: Status 404 returned error can't find the container with id 334450f2c8760f6a70e4305009afa9861d4386fd53ac4de80a6609df962e6090 Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.451775 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc"] Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.902538 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" event={"ID":"c5e5ab85-6606-4e44-bc7b-8fe2df836f49","Type":"ContainerStarted","Data":"c5ff730993232a7907cc227d5d3f9c00741a0c4ae8c9f603c48afdd77e50baac"} Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.904296 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lg5x2" event={"ID":"34af49c2-e879-487f-b92e-c178ea679f4d","Type":"ContainerStarted","Data":"c8ce0091e885b8d468b4b1ad7b7736a1d4a264646af1868f0cb911f45f7b8b39"} Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.905428 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" event={"ID":"d3b269c8-9fb0-4982-a512-eb5e91d37a97","Type":"ContainerStarted","Data":"446b250283e587078768e4de9bdc972343f5d481511efa03e4feeffee4794a38"} Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.907024 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-db8dd7775-q6mmf" event={"ID":"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87","Type":"ContainerStarted","Data":"443f4066ac515317024ca06f821383213895dec3baf8d8a06855b282a3d5472f"} Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.907055 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-db8dd7775-q6mmf" event={"ID":"ad9b84c3-2c65-4b1d-a1b7-96d3ef6bce87","Type":"ContainerStarted","Data":"334450f2c8760f6a70e4305009afa9861d4386fd53ac4de80a6609df962e6090"} Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.909373 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" event={"ID":"c7329403-a7f9-45bb-93e9-7fdc91953d06","Type":"ContainerStarted","Data":"df23d628fe4f742c7e5a5be48b5991b4428910a97e86c006b73991c8f699ac39"} Oct 10 14:58:36 crc kubenswrapper[4788]: I1010 14:58:36.929094 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-db8dd7775-q6mmf" podStartSLOduration=1.929070676 podStartE2EDuration="1.929070676s" podCreationTimestamp="2025-10-10 14:58:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 14:58:36.924695584 +0000 UTC m=+819.374411132" watchObservedRunningTime="2025-10-10 14:58:36.929070676 +0000 UTC m=+819.378786224" Oct 10 14:58:37 crc kubenswrapper[4788]: I1010 14:58:37.622566 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mfhz2"] Oct 10 14:58:37 crc kubenswrapper[4788]: I1010 14:58:37.622793 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mfhz2" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" containerName="registry-server" containerID="cri-o://1b68642f9c8585b60b7f3c1afd9c5cba67cbfccdc8d3be7e5de8c9bc8555251e" gracePeriod=2 Oct 10 14:58:37 crc kubenswrapper[4788]: I1010 14:58:37.920218 4788 generic.go:334] "Generic (PLEG): container finished" podID="15beebd9-6640-4117-ab2a-047a42b83a79" containerID="1b68642f9c8585b60b7f3c1afd9c5cba67cbfccdc8d3be7e5de8c9bc8555251e" exitCode=0 Oct 10 14:58:37 crc kubenswrapper[4788]: I1010 14:58:37.921117 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfhz2" event={"ID":"15beebd9-6640-4117-ab2a-047a42b83a79","Type":"ContainerDied","Data":"1b68642f9c8585b60b7f3c1afd9c5cba67cbfccdc8d3be7e5de8c9bc8555251e"} Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.617026 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.703120 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc8hc\" (UniqueName: \"kubernetes.io/projected/15beebd9-6640-4117-ab2a-047a42b83a79-kube-api-access-cc8hc\") pod \"15beebd9-6640-4117-ab2a-047a42b83a79\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.703523 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-catalog-content\") pod \"15beebd9-6640-4117-ab2a-047a42b83a79\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.703834 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-utilities\") pod \"15beebd9-6640-4117-ab2a-047a42b83a79\" (UID: \"15beebd9-6640-4117-ab2a-047a42b83a79\") " Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.705061 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-utilities" (OuterVolumeSpecName: "utilities") pod "15beebd9-6640-4117-ab2a-047a42b83a79" (UID: "15beebd9-6640-4117-ab2a-047a42b83a79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.729611 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15beebd9-6640-4117-ab2a-047a42b83a79-kube-api-access-cc8hc" (OuterVolumeSpecName: "kube-api-access-cc8hc") pod "15beebd9-6640-4117-ab2a-047a42b83a79" (UID: "15beebd9-6640-4117-ab2a-047a42b83a79"). InnerVolumeSpecName "kube-api-access-cc8hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.769226 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15beebd9-6640-4117-ab2a-047a42b83a79" (UID: "15beebd9-6640-4117-ab2a-047a42b83a79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.805975 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.806013 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15beebd9-6640-4117-ab2a-047a42b83a79-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.806025 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc8hc\" (UniqueName: \"kubernetes.io/projected/15beebd9-6640-4117-ab2a-047a42b83a79-kube-api-access-cc8hc\") on node \"crc\" DevicePath \"\"" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.938733 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" event={"ID":"d3b269c8-9fb0-4982-a512-eb5e91d37a97","Type":"ContainerStarted","Data":"cabacb856fe84b98adb3509c39a9daefc6a4ad7de1168800c7b237f3891f2542"} Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.939080 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.941505 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" event={"ID":"c7329403-a7f9-45bb-93e9-7fdc91953d06","Type":"ContainerStarted","Data":"d29adbf223862eb3b53303a79cb46329cb796727761254e30858aacc3a0161e6"} Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.947528 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfhz2" event={"ID":"15beebd9-6640-4117-ab2a-047a42b83a79","Type":"ContainerDied","Data":"396ea2658e1ec1961fb6a12244ab83b8afaa081baa92c5934c0380e650fa4c61"} Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.947573 4788 scope.go:117] "RemoveContainer" containerID="1b68642f9c8585b60b7f3c1afd9c5cba67cbfccdc8d3be7e5de8c9bc8555251e" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.947712 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mfhz2" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.992948 4788 scope.go:117] "RemoveContainer" containerID="60016f0d06bc1b87e425bc41d54071788445976c28e369c57c3206db42e399ec" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.995078 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" podStartSLOduration=1.410068361 podStartE2EDuration="3.995060541s" podCreationTimestamp="2025-10-10 14:58:35 +0000 UTC" firstStartedPulling="2025-10-10 14:58:36.102563354 +0000 UTC m=+818.552278902" lastFinishedPulling="2025-10-10 14:58:38.687555524 +0000 UTC m=+821.137271082" observedRunningTime="2025-10-10 14:58:38.967591794 +0000 UTC m=+821.417307342" watchObservedRunningTime="2025-10-10 14:58:38.995060541 +0000 UTC m=+821.444776089" Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.997303 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mfhz2"] Oct 10 14:58:38 crc kubenswrapper[4788]: I1010 14:58:38.999728 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mfhz2"] Oct 10 14:58:39 crc kubenswrapper[4788]: I1010 14:58:39.018246 4788 scope.go:117] "RemoveContainer" containerID="06ef843888d1cd292ed6c13eef398242e61ee630baf38acfa7fb7ac7aa461c92" Oct 10 14:58:39 crc kubenswrapper[4788]: I1010 14:58:39.960539 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" event={"ID":"c5e5ab85-6606-4e44-bc7b-8fe2df836f49","Type":"ContainerStarted","Data":"ad4a9721a5adc4f72f66c45e0ac23b137d93172d874b99d5495c55274b714966"} Oct 10 14:58:39 crc kubenswrapper[4788]: I1010 14:58:39.965839 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lg5x2" event={"ID":"34af49c2-e879-487f-b92e-c178ea679f4d","Type":"ContainerStarted","Data":"e3c7f8f69fd216e272ed4af8ef4b46986d9fc3db28ef8d2d207f5f717572710a"} Oct 10 14:58:39 crc kubenswrapper[4788]: I1010 14:58:39.966130 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:39 crc kubenswrapper[4788]: I1010 14:58:39.984060 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-d78sc" podStartSLOduration=1.820818013 podStartE2EDuration="4.98403921s" podCreationTimestamp="2025-10-10 14:58:35 +0000 UTC" firstStartedPulling="2025-10-10 14:58:36.491364552 +0000 UTC m=+818.941080100" lastFinishedPulling="2025-10-10 14:58:39.654585749 +0000 UTC m=+822.104301297" observedRunningTime="2025-10-10 14:58:39.982307831 +0000 UTC m=+822.432023379" watchObservedRunningTime="2025-10-10 14:58:39.98403921 +0000 UTC m=+822.433754758" Oct 10 14:58:40 crc kubenswrapper[4788]: I1010 14:58:40.017582 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-lg5x2" podStartSLOduration=2.257192258 podStartE2EDuration="5.017546655s" podCreationTimestamp="2025-10-10 14:58:35 +0000 UTC" firstStartedPulling="2025-10-10 14:58:35.927453714 +0000 UTC m=+818.377169252" lastFinishedPulling="2025-10-10 14:58:38.687808101 +0000 UTC m=+821.137523649" observedRunningTime="2025-10-10 14:58:40.009891311 +0000 UTC m=+822.459606859" watchObservedRunningTime="2025-10-10 14:58:40.017546655 +0000 UTC m=+822.467262233" Oct 10 14:58:40 crc kubenswrapper[4788]: I1010 14:58:40.242728 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" path="/var/lib/kubelet/pods/15beebd9-6640-4117-ab2a-047a42b83a79/volumes" Oct 10 14:58:41 crc kubenswrapper[4788]: I1010 14:58:41.984561 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" event={"ID":"c7329403-a7f9-45bb-93e9-7fdc91953d06","Type":"ContainerStarted","Data":"a5579b5ec6ec54427a6db5a32e0bb31a9def72274448942fe8fde6ebe033b971"} Oct 10 14:58:42 crc kubenswrapper[4788]: I1010 14:58:42.015370 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-7crrv" podStartSLOduration=1.9026555790000002 podStartE2EDuration="7.015345086s" podCreationTimestamp="2025-10-10 14:58:35 +0000 UTC" firstStartedPulling="2025-10-10 14:58:36.171281624 +0000 UTC m=+818.620997172" lastFinishedPulling="2025-10-10 14:58:41.283971141 +0000 UTC m=+823.733686679" observedRunningTime="2025-10-10 14:58:42.008042202 +0000 UTC m=+824.457757780" watchObservedRunningTime="2025-10-10 14:58:42.015345086 +0000 UTC m=+824.465060644" Oct 10 14:58:45 crc kubenswrapper[4788]: I1010 14:58:45.886027 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-lg5x2" Oct 10 14:58:46 crc kubenswrapper[4788]: I1010 14:58:46.213775 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:46 crc kubenswrapper[4788]: I1010 14:58:46.213822 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:46 crc kubenswrapper[4788]: I1010 14:58:46.219198 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:47 crc kubenswrapper[4788]: I1010 14:58:47.030035 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-db8dd7775-q6mmf" Oct 10 14:58:47 crc kubenswrapper[4788]: I1010 14:58:47.105113 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mgbhm"] Oct 10 14:58:55 crc kubenswrapper[4788]: I1010 14:58:55.846008 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5q4md" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.286215 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4"] Oct 10 14:59:11 crc kubenswrapper[4788]: E1010 14:59:11.287349 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" containerName="registry-server" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.287370 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" containerName="registry-server" Oct 10 14:59:11 crc kubenswrapper[4788]: E1010 14:59:11.287403 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" containerName="extract-utilities" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.287411 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" containerName="extract-utilities" Oct 10 14:59:11 crc kubenswrapper[4788]: E1010 14:59:11.287426 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" containerName="extract-content" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.287461 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" containerName="extract-content" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.287591 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="15beebd9-6640-4117-ab2a-047a42b83a79" containerName="registry-server" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.288708 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.291458 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.299568 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4"] Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.488124 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.488285 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.488483 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w7pj\" (UniqueName: \"kubernetes.io/projected/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-kube-api-access-9w7pj\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.589125 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w7pj\" (UniqueName: \"kubernetes.io/projected/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-kube-api-access-9w7pj\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.589233 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.589267 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.589933 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.589976 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.617419 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w7pj\" (UniqueName: \"kubernetes.io/projected/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-kube-api-access-9w7pj\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:11 crc kubenswrapper[4788]: I1010 14:59:11.640783 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.077467 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4"] Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.153786 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-mgbhm" podUID="d8826b11-1579-4653-953e-0895bc1d16b5" containerName="console" containerID="cri-o://338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc" gracePeriod=15 Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.193775 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" event={"ID":"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9","Type":"ContainerStarted","Data":"5d4cce1c191b7da91db6ea190ebcfbf1041ddac8ee9771d297d71682092802b9"} Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.608955 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mgbhm_d8826b11-1579-4653-953e-0895bc1d16b5/console/0.log" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.609516 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.719793 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-console-config\") pod \"d8826b11-1579-4653-953e-0895bc1d16b5\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.719885 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-oauth-serving-cert\") pod \"d8826b11-1579-4653-953e-0895bc1d16b5\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.719937 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-service-ca\") pod \"d8826b11-1579-4653-953e-0895bc1d16b5\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.719969 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-trusted-ca-bundle\") pod \"d8826b11-1579-4653-953e-0895bc1d16b5\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.720014 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv7vk\" (UniqueName: \"kubernetes.io/projected/d8826b11-1579-4653-953e-0895bc1d16b5-kube-api-access-kv7vk\") pod \"d8826b11-1579-4653-953e-0895bc1d16b5\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.720094 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-serving-cert\") pod \"d8826b11-1579-4653-953e-0895bc1d16b5\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.720201 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-oauth-config\") pod \"d8826b11-1579-4653-953e-0895bc1d16b5\" (UID: \"d8826b11-1579-4653-953e-0895bc1d16b5\") " Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.721670 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-console-config" (OuterVolumeSpecName: "console-config") pod "d8826b11-1579-4653-953e-0895bc1d16b5" (UID: "d8826b11-1579-4653-953e-0895bc1d16b5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.721664 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-service-ca" (OuterVolumeSpecName: "service-ca") pod "d8826b11-1579-4653-953e-0895bc1d16b5" (UID: "d8826b11-1579-4653-953e-0895bc1d16b5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.721715 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d8826b11-1579-4653-953e-0895bc1d16b5" (UID: "d8826b11-1579-4653-953e-0895bc1d16b5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.721868 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d8826b11-1579-4653-953e-0895bc1d16b5" (UID: "d8826b11-1579-4653-953e-0895bc1d16b5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.728801 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d8826b11-1579-4653-953e-0895bc1d16b5" (UID: "d8826b11-1579-4653-953e-0895bc1d16b5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.729088 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d8826b11-1579-4653-953e-0895bc1d16b5" (UID: "d8826b11-1579-4653-953e-0895bc1d16b5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.729231 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8826b11-1579-4653-953e-0895bc1d16b5-kube-api-access-kv7vk" (OuterVolumeSpecName: "kube-api-access-kv7vk") pod "d8826b11-1579-4653-953e-0895bc1d16b5" (UID: "d8826b11-1579-4653-953e-0895bc1d16b5"). InnerVolumeSpecName "kube-api-access-kv7vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.823215 4788 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.823286 4788 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-console-config\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.823308 4788 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.823326 4788 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.823345 4788 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8826b11-1579-4653-953e-0895bc1d16b5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.823369 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv7vk\" (UniqueName: \"kubernetes.io/projected/d8826b11-1579-4653-953e-0895bc1d16b5-kube-api-access-kv7vk\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:12 crc kubenswrapper[4788]: I1010 14:59:12.823404 4788 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8826b11-1579-4653-953e-0895bc1d16b5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.205439 4788 generic.go:334] "Generic (PLEG): container finished" podID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerID="5fbe5a907ec0c35d3e1119ac130d242494f0d9fb62724df99865b2416f54bdc4" exitCode=0 Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.205562 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" event={"ID":"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9","Type":"ContainerDied","Data":"5fbe5a907ec0c35d3e1119ac130d242494f0d9fb62724df99865b2416f54bdc4"} Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.209840 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mgbhm_d8826b11-1579-4653-953e-0895bc1d16b5/console/0.log" Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.209912 4788 generic.go:334] "Generic (PLEG): container finished" podID="d8826b11-1579-4653-953e-0895bc1d16b5" containerID="338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc" exitCode=2 Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.209959 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgbhm" event={"ID":"d8826b11-1579-4653-953e-0895bc1d16b5","Type":"ContainerDied","Data":"338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc"} Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.210001 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgbhm" event={"ID":"d8826b11-1579-4653-953e-0895bc1d16b5","Type":"ContainerDied","Data":"92e63c9af9921dd9cb2e51ee57fa4720cf93e2197e29970899340b7861883b01"} Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.210028 4788 scope.go:117] "RemoveContainer" containerID="338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc" Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.210232 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgbhm" Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.271353 4788 scope.go:117] "RemoveContainer" containerID="338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc" Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.271523 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mgbhm"] Oct 10 14:59:13 crc kubenswrapper[4788]: E1010 14:59:13.272440 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc\": container with ID starting with 338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc not found: ID does not exist" containerID="338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc" Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.272506 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc"} err="failed to get container status \"338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc\": rpc error: code = NotFound desc = could not find container \"338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc\": container with ID starting with 338adfdc1f6824941ece18c7ab6cf4be540617a14b0fc50e8ba9541fbb15acfc not found: ID does not exist" Oct 10 14:59:13 crc kubenswrapper[4788]: I1010 14:59:13.279009 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-mgbhm"] Oct 10 14:59:14 crc kubenswrapper[4788]: I1010 14:59:14.243374 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8826b11-1579-4653-953e-0895bc1d16b5" path="/var/lib/kubelet/pods/d8826b11-1579-4653-953e-0895bc1d16b5/volumes" Oct 10 14:59:15 crc kubenswrapper[4788]: I1010 14:59:15.231338 4788 generic.go:334] "Generic (PLEG): container finished" podID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerID="68898ce7b95b7735b721f9b4ef59da13cff2ba9ca76fb357773fbfa6b8b73c02" exitCode=0 Oct 10 14:59:15 crc kubenswrapper[4788]: I1010 14:59:15.231454 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" event={"ID":"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9","Type":"ContainerDied","Data":"68898ce7b95b7735b721f9b4ef59da13cff2ba9ca76fb357773fbfa6b8b73c02"} Oct 10 14:59:16 crc kubenswrapper[4788]: I1010 14:59:16.244013 4788 generic.go:334] "Generic (PLEG): container finished" podID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerID="590397119d86bfc3ff18b9b43eb2829356bf954378b0a0b5c252228699edc41c" exitCode=0 Oct 10 14:59:16 crc kubenswrapper[4788]: I1010 14:59:16.245351 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" event={"ID":"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9","Type":"ContainerDied","Data":"590397119d86bfc3ff18b9b43eb2829356bf954378b0a0b5c252228699edc41c"} Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.596769 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.697830 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-util\") pod \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.697912 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-bundle\") pod \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.697979 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w7pj\" (UniqueName: \"kubernetes.io/projected/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-kube-api-access-9w7pj\") pod \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\" (UID: \"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9\") " Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.699429 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-bundle" (OuterVolumeSpecName: "bundle") pod "60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" (UID: "60d1e35d-ed52-4de4-8aef-b10d13e2d5c9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.705661 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-kube-api-access-9w7pj" (OuterVolumeSpecName: "kube-api-access-9w7pj") pod "60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" (UID: "60d1e35d-ed52-4de4-8aef-b10d13e2d5c9"). InnerVolumeSpecName "kube-api-access-9w7pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.799671 4788 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.799712 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w7pj\" (UniqueName: \"kubernetes.io/projected/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-kube-api-access-9w7pj\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:17 crc kubenswrapper[4788]: I1010 14:59:17.976508 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-util" (OuterVolumeSpecName: "util") pod "60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" (UID: "60d1e35d-ed52-4de4-8aef-b10d13e2d5c9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:59:18 crc kubenswrapper[4788]: I1010 14:59:18.001657 4788 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60d1e35d-ed52-4de4-8aef-b10d13e2d5c9-util\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:18 crc kubenswrapper[4788]: I1010 14:59:18.265581 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" event={"ID":"60d1e35d-ed52-4de4-8aef-b10d13e2d5c9","Type":"ContainerDied","Data":"5d4cce1c191b7da91db6ea190ebcfbf1041ddac8ee9771d297d71682092802b9"} Oct 10 14:59:18 crc kubenswrapper[4788]: I1010 14:59:18.265676 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d4cce1c191b7da91db6ea190ebcfbf1041ddac8ee9771d297d71682092802b9" Oct 10 14:59:18 crc kubenswrapper[4788]: I1010 14:59:18.265827 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.428644 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v7nkm"] Oct 10 14:59:21 crc kubenswrapper[4788]: E1010 14:59:21.429678 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerName="pull" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.429696 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerName="pull" Oct 10 14:59:21 crc kubenswrapper[4788]: E1010 14:59:21.429707 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerName="util" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.429716 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerName="util" Oct 10 14:59:21 crc kubenswrapper[4788]: E1010 14:59:21.429728 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerName="extract" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.429736 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerName="extract" Oct 10 14:59:21 crc kubenswrapper[4788]: E1010 14:59:21.429748 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8826b11-1579-4653-953e-0895bc1d16b5" containerName="console" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.429755 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8826b11-1579-4653-953e-0895bc1d16b5" containerName="console" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.429879 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8826b11-1579-4653-953e-0895bc1d16b5" containerName="console" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.429894 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="60d1e35d-ed52-4de4-8aef-b10d13e2d5c9" containerName="extract" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.431191 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.449819 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v7nkm"] Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.554429 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-catalog-content\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.554885 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-utilities\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.554989 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvfht\" (UniqueName: \"kubernetes.io/projected/19d21c15-3465-4f9b-8f59-5255180f9391-kube-api-access-bvfht\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.657127 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-catalog-content\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.657483 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-utilities\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.657563 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvfht\" (UniqueName: \"kubernetes.io/projected/19d21c15-3465-4f9b-8f59-5255180f9391-kube-api-access-bvfht\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.657689 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-catalog-content\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.657951 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-utilities\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.685185 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvfht\" (UniqueName: \"kubernetes.io/projected/19d21c15-3465-4f9b-8f59-5255180f9391-kube-api-access-bvfht\") pod \"certified-operators-v7nkm\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:21 crc kubenswrapper[4788]: I1010 14:59:21.776830 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:22 crc kubenswrapper[4788]: I1010 14:59:22.035393 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v7nkm"] Oct 10 14:59:22 crc kubenswrapper[4788]: I1010 14:59:22.301884 4788 generic.go:334] "Generic (PLEG): container finished" podID="19d21c15-3465-4f9b-8f59-5255180f9391" containerID="9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc" exitCode=0 Oct 10 14:59:22 crc kubenswrapper[4788]: I1010 14:59:22.302027 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7nkm" event={"ID":"19d21c15-3465-4f9b-8f59-5255180f9391","Type":"ContainerDied","Data":"9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc"} Oct 10 14:59:22 crc kubenswrapper[4788]: I1010 14:59:22.302409 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7nkm" event={"ID":"19d21c15-3465-4f9b-8f59-5255180f9391","Type":"ContainerStarted","Data":"3e338c12025872a9fe14d73fa12a4e89a39f7e2c89c69a444663d41931709986"} Oct 10 14:59:23 crc kubenswrapper[4788]: I1010 14:59:23.309484 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7nkm" event={"ID":"19d21c15-3465-4f9b-8f59-5255180f9391","Type":"ContainerStarted","Data":"30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c"} Oct 10 14:59:24 crc kubenswrapper[4788]: I1010 14:59:24.322813 4788 generic.go:334] "Generic (PLEG): container finished" podID="19d21c15-3465-4f9b-8f59-5255180f9391" containerID="30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c" exitCode=0 Oct 10 14:59:24 crc kubenswrapper[4788]: I1010 14:59:24.322871 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7nkm" event={"ID":"19d21c15-3465-4f9b-8f59-5255180f9391","Type":"ContainerDied","Data":"30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c"} Oct 10 14:59:25 crc kubenswrapper[4788]: I1010 14:59:25.343958 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7nkm" event={"ID":"19d21c15-3465-4f9b-8f59-5255180f9391","Type":"ContainerStarted","Data":"6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875"} Oct 10 14:59:25 crc kubenswrapper[4788]: I1010 14:59:25.364985 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v7nkm" podStartSLOduration=1.93879186 podStartE2EDuration="4.364958226s" podCreationTimestamp="2025-10-10 14:59:21 +0000 UTC" firstStartedPulling="2025-10-10 14:59:22.30380291 +0000 UTC m=+864.753518458" lastFinishedPulling="2025-10-10 14:59:24.729969276 +0000 UTC m=+867.179684824" observedRunningTime="2025-10-10 14:59:25.363708522 +0000 UTC m=+867.813424070" watchObservedRunningTime="2025-10-10 14:59:25.364958226 +0000 UTC m=+867.814673774" Oct 10 14:59:27 crc kubenswrapper[4788]: I1010 14:59:27.995996 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq"] Oct 10 14:59:27 crc kubenswrapper[4788]: I1010 14:59:27.996949 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.000843 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.000857 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.001096 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.001413 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-x5xlq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.001658 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.016324 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq"] Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.151803 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68109e8b-88f9-48ad-b162-282574174516-apiservice-cert\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.151870 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68109e8b-88f9-48ad-b162-282574174516-webhook-cert\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.151891 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k7fv\" (UniqueName: \"kubernetes.io/projected/68109e8b-88f9-48ad-b162-282574174516-kube-api-access-7k7fv\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.254168 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68109e8b-88f9-48ad-b162-282574174516-apiservice-cert\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.254240 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68109e8b-88f9-48ad-b162-282574174516-webhook-cert\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.254269 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k7fv\" (UniqueName: \"kubernetes.io/projected/68109e8b-88f9-48ad-b162-282574174516-kube-api-access-7k7fv\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.262030 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68109e8b-88f9-48ad-b162-282574174516-webhook-cert\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.262094 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68109e8b-88f9-48ad-b162-282574174516-apiservice-cert\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.272675 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k7fv\" (UniqueName: \"kubernetes.io/projected/68109e8b-88f9-48ad-b162-282574174516-kube-api-access-7k7fv\") pod \"metallb-operator-controller-manager-5b8b68658d-sn5zq\" (UID: \"68109e8b-88f9-48ad-b162-282574174516\") " pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.314427 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.348025 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x"] Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.348909 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.355509 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.355763 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pxg8c" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.355914 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.364681 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x"] Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.460476 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f9b7183-67aa-4cea-83c6-59a32ff3f027-webhook-cert\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.460551 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h4l5\" (UniqueName: \"kubernetes.io/projected/2f9b7183-67aa-4cea-83c6-59a32ff3f027-kube-api-access-2h4l5\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.460588 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f9b7183-67aa-4cea-83c6-59a32ff3f027-apiservice-cert\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.561748 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f9b7183-67aa-4cea-83c6-59a32ff3f027-webhook-cert\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.561842 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h4l5\" (UniqueName: \"kubernetes.io/projected/2f9b7183-67aa-4cea-83c6-59a32ff3f027-kube-api-access-2h4l5\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.561878 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f9b7183-67aa-4cea-83c6-59a32ff3f027-apiservice-cert\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.573123 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2f9b7183-67aa-4cea-83c6-59a32ff3f027-apiservice-cert\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.573156 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2f9b7183-67aa-4cea-83c6-59a32ff3f027-webhook-cert\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.611031 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h4l5\" (UniqueName: \"kubernetes.io/projected/2f9b7183-67aa-4cea-83c6-59a32ff3f027-kube-api-access-2h4l5\") pod \"metallb-operator-webhook-server-d8576d6b4-hr28x\" (UID: \"2f9b7183-67aa-4cea-83c6-59a32ff3f027\") " pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.673021 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq"] Oct 10 14:59:28 crc kubenswrapper[4788]: W1010 14:59:28.701318 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68109e8b_88f9_48ad_b162_282574174516.slice/crio-359746426a3310e7eef57664530a40b6d03cc16e8fa1181634914f5cb5df9270 WatchSource:0}: Error finding container 359746426a3310e7eef57664530a40b6d03cc16e8fa1181634914f5cb5df9270: Status 404 returned error can't find the container with id 359746426a3310e7eef57664530a40b6d03cc16e8fa1181634914f5cb5df9270 Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.706856 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:28 crc kubenswrapper[4788]: I1010 14:59:28.972705 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x"] Oct 10 14:59:29 crc kubenswrapper[4788]: I1010 14:59:29.398911 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" event={"ID":"68109e8b-88f9-48ad-b162-282574174516","Type":"ContainerStarted","Data":"359746426a3310e7eef57664530a40b6d03cc16e8fa1181634914f5cb5df9270"} Oct 10 14:59:29 crc kubenswrapper[4788]: I1010 14:59:29.401603 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" event={"ID":"2f9b7183-67aa-4cea-83c6-59a32ff3f027","Type":"ContainerStarted","Data":"6f6f0bb7d20d6231fbae4868a5ca48395e4eb07cc5437d6f3b6408803c63c223"} Oct 10 14:59:29 crc kubenswrapper[4788]: I1010 14:59:29.406599 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:59:29 crc kubenswrapper[4788]: I1010 14:59:29.406670 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 14:59:31 crc kubenswrapper[4788]: I1010 14:59:31.777175 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:31 crc kubenswrapper[4788]: I1010 14:59:31.777499 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:31 crc kubenswrapper[4788]: I1010 14:59:31.820460 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:32 crc kubenswrapper[4788]: I1010 14:59:32.528260 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:33 crc kubenswrapper[4788]: I1010 14:59:33.822928 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v7nkm"] Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.478218 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" event={"ID":"68109e8b-88f9-48ad-b162-282574174516","Type":"ContainerStarted","Data":"95d12a4b1a05ee7a4a90987bb4785f38c3c6233a4aa201c63776beafa6f51639"} Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.479948 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.483429 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" event={"ID":"2f9b7183-67aa-4cea-83c6-59a32ff3f027","Type":"ContainerStarted","Data":"6cb42401a93c8a3ed59972b957aebd871e1484edd02c5e4e8e18cfa7cfd11692"} Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.483459 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v7nkm" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" containerName="registry-server" containerID="cri-o://6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875" gracePeriod=2 Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.483707 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.533217 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" podStartSLOduration=2.256415507 podStartE2EDuration="7.533195343s" podCreationTimestamp="2025-10-10 14:59:27 +0000 UTC" firstStartedPulling="2025-10-10 14:59:28.711308197 +0000 UTC m=+871.161023735" lastFinishedPulling="2025-10-10 14:59:33.988087973 +0000 UTC m=+876.437803571" observedRunningTime="2025-10-10 14:59:34.50250185 +0000 UTC m=+876.952217398" watchObservedRunningTime="2025-10-10 14:59:34.533195343 +0000 UTC m=+876.982910891" Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.535100 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" podStartSLOduration=1.505545433 podStartE2EDuration="6.535079615s" podCreationTimestamp="2025-10-10 14:59:28 +0000 UTC" firstStartedPulling="2025-10-10 14:59:28.983353674 +0000 UTC m=+871.433069222" lastFinishedPulling="2025-10-10 14:59:34.012887846 +0000 UTC m=+876.462603404" observedRunningTime="2025-10-10 14:59:34.530540211 +0000 UTC m=+876.980255779" watchObservedRunningTime="2025-10-10 14:59:34.535079615 +0000 UTC m=+876.984795163" Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.935344 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:34 crc kubenswrapper[4788]: I1010 14:59:34.999742 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-utilities\") pod \"19d21c15-3465-4f9b-8f59-5255180f9391\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:34.999892 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvfht\" (UniqueName: \"kubernetes.io/projected/19d21c15-3465-4f9b-8f59-5255180f9391-kube-api-access-bvfht\") pod \"19d21c15-3465-4f9b-8f59-5255180f9391\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:34.999992 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-catalog-content\") pod \"19d21c15-3465-4f9b-8f59-5255180f9391\" (UID: \"19d21c15-3465-4f9b-8f59-5255180f9391\") " Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.005552 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-utilities" (OuterVolumeSpecName: "utilities") pod "19d21c15-3465-4f9b-8f59-5255180f9391" (UID: "19d21c15-3465-4f9b-8f59-5255180f9391"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.011376 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d21c15-3465-4f9b-8f59-5255180f9391-kube-api-access-bvfht" (OuterVolumeSpecName: "kube-api-access-bvfht") pod "19d21c15-3465-4f9b-8f59-5255180f9391" (UID: "19d21c15-3465-4f9b-8f59-5255180f9391"). InnerVolumeSpecName "kube-api-access-bvfht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.067319 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19d21c15-3465-4f9b-8f59-5255180f9391" (UID: "19d21c15-3465-4f9b-8f59-5255180f9391"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.100776 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.100810 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d21c15-3465-4f9b-8f59-5255180f9391-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.100820 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvfht\" (UniqueName: \"kubernetes.io/projected/19d21c15-3465-4f9b-8f59-5255180f9391-kube-api-access-bvfht\") on node \"crc\" DevicePath \"\"" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.491794 4788 generic.go:334] "Generic (PLEG): container finished" podID="19d21c15-3465-4f9b-8f59-5255180f9391" containerID="6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875" exitCode=0 Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.491852 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7nkm" event={"ID":"19d21c15-3465-4f9b-8f59-5255180f9391","Type":"ContainerDied","Data":"6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875"} Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.492604 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v7nkm" event={"ID":"19d21c15-3465-4f9b-8f59-5255180f9391","Type":"ContainerDied","Data":"3e338c12025872a9fe14d73fa12a4e89a39f7e2c89c69a444663d41931709986"} Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.492656 4788 scope.go:117] "RemoveContainer" containerID="6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.491937 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v7nkm" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.524072 4788 scope.go:117] "RemoveContainer" containerID="30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.529288 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v7nkm"] Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.539921 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v7nkm"] Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.571889 4788 scope.go:117] "RemoveContainer" containerID="9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.591195 4788 scope.go:117] "RemoveContainer" containerID="6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875" Oct 10 14:59:35 crc kubenswrapper[4788]: E1010 14:59:35.591691 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875\": container with ID starting with 6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875 not found: ID does not exist" containerID="6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.591737 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875"} err="failed to get container status \"6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875\": rpc error: code = NotFound desc = could not find container \"6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875\": container with ID starting with 6c660492956d159a76db5126fd9d90a9b11dfd6d19dce72ec46874f698e97875 not found: ID does not exist" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.591811 4788 scope.go:117] "RemoveContainer" containerID="30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c" Oct 10 14:59:35 crc kubenswrapper[4788]: E1010 14:59:35.592181 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c\": container with ID starting with 30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c not found: ID does not exist" containerID="30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.592217 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c"} err="failed to get container status \"30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c\": rpc error: code = NotFound desc = could not find container \"30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c\": container with ID starting with 30261e56a93c7a3dce86ee4826cff2dab901420c7724928aaeacb09eea9d402c not found: ID does not exist" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.592245 4788 scope.go:117] "RemoveContainer" containerID="9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc" Oct 10 14:59:35 crc kubenswrapper[4788]: E1010 14:59:35.592713 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc\": container with ID starting with 9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc not found: ID does not exist" containerID="9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc" Oct 10 14:59:35 crc kubenswrapper[4788]: I1010 14:59:35.592749 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc"} err="failed to get container status \"9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc\": rpc error: code = NotFound desc = could not find container \"9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc\": container with ID starting with 9c6a3cf936e4bee1b0e7cd42fe13c99ea343f7c1b8425452f6810ee2a8221dfc not found: ID does not exist" Oct 10 14:59:36 crc kubenswrapper[4788]: I1010 14:59:36.244273 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" path="/var/lib/kubelet/pods/19d21c15-3465-4f9b-8f59-5255180f9391/volumes" Oct 10 14:59:48 crc kubenswrapper[4788]: I1010 14:59:48.711803 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-d8576d6b4-hr28x" Oct 10 14:59:59 crc kubenswrapper[4788]: I1010 14:59:59.406789 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 14:59:59 crc kubenswrapper[4788]: I1010 14:59:59.407776 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.158767 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn"] Oct 10 15:00:00 crc kubenswrapper[4788]: E1010 15:00:00.159652 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" containerName="extract-content" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.159675 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" containerName="extract-content" Oct 10 15:00:00 crc kubenswrapper[4788]: E1010 15:00:00.159692 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" containerName="extract-utilities" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.159702 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" containerName="extract-utilities" Oct 10 15:00:00 crc kubenswrapper[4788]: E1010 15:00:00.159718 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" containerName="registry-server" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.159727 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" containerName="registry-server" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.159853 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d21c15-3465-4f9b-8f59-5255180f9391" containerName="registry-server" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.160571 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.212371 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.212586 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.255301 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn"] Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.267395 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-config-volume\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.267475 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm2rj\" (UniqueName: \"kubernetes.io/projected/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-kube-api-access-wm2rj\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.267720 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-secret-volume\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.368937 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-secret-volume\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.369015 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-config-volume\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.369928 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm2rj\" (UniqueName: \"kubernetes.io/projected/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-kube-api-access-wm2rj\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.370342 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-config-volume\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.376357 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-secret-volume\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.387949 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm2rj\" (UniqueName: \"kubernetes.io/projected/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-kube-api-access-wm2rj\") pod \"collect-profiles-29335140-89dcn\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.548127 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:00 crc kubenswrapper[4788]: I1010 15:00:00.986233 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn"] Oct 10 15:00:00 crc kubenswrapper[4788]: W1010 15:00:00.996582 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a3c9bad_9fee_4dd0_b553_52cebeb42d12.slice/crio-8c2f94b5eacd9b20217771c63cdb899a5ce18dbc2e034a1009a87c031c1625b6 WatchSource:0}: Error finding container 8c2f94b5eacd9b20217771c63cdb899a5ce18dbc2e034a1009a87c031c1625b6: Status 404 returned error can't find the container with id 8c2f94b5eacd9b20217771c63cdb899a5ce18dbc2e034a1009a87c031c1625b6 Oct 10 15:00:01 crc kubenswrapper[4788]: I1010 15:00:01.665817 4788 generic.go:334] "Generic (PLEG): container finished" podID="9a3c9bad-9fee-4dd0-b553-52cebeb42d12" containerID="a1adbcca9500ed3d840078abe319997f98e311b25bf852506f5584458cc7ef55" exitCode=0 Oct 10 15:00:01 crc kubenswrapper[4788]: I1010 15:00:01.665919 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" event={"ID":"9a3c9bad-9fee-4dd0-b553-52cebeb42d12","Type":"ContainerDied","Data":"a1adbcca9500ed3d840078abe319997f98e311b25bf852506f5584458cc7ef55"} Oct 10 15:00:01 crc kubenswrapper[4788]: I1010 15:00:01.667691 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" event={"ID":"9a3c9bad-9fee-4dd0-b553-52cebeb42d12","Type":"ContainerStarted","Data":"8c2f94b5eacd9b20217771c63cdb899a5ce18dbc2e034a1009a87c031c1625b6"} Oct 10 15:00:02 crc kubenswrapper[4788]: I1010 15:00:02.977356 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.108924 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-secret-volume\") pod \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.109018 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm2rj\" (UniqueName: \"kubernetes.io/projected/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-kube-api-access-wm2rj\") pod \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.109048 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-config-volume\") pod \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\" (UID: \"9a3c9bad-9fee-4dd0-b553-52cebeb42d12\") " Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.109793 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-config-volume" (OuterVolumeSpecName: "config-volume") pod "9a3c9bad-9fee-4dd0-b553-52cebeb42d12" (UID: "9a3c9bad-9fee-4dd0-b553-52cebeb42d12"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.114091 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9a3c9bad-9fee-4dd0-b553-52cebeb42d12" (UID: "9a3c9bad-9fee-4dd0-b553-52cebeb42d12"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.114342 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-kube-api-access-wm2rj" (OuterVolumeSpecName: "kube-api-access-wm2rj") pod "9a3c9bad-9fee-4dd0-b553-52cebeb42d12" (UID: "9a3c9bad-9fee-4dd0-b553-52cebeb42d12"). InnerVolumeSpecName "kube-api-access-wm2rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.211022 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.211089 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm2rj\" (UniqueName: \"kubernetes.io/projected/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-kube-api-access-wm2rj\") on node \"crc\" DevicePath \"\"" Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.211104 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a3c9bad-9fee-4dd0-b553-52cebeb42d12-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.681084 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" event={"ID":"9a3c9bad-9fee-4dd0-b553-52cebeb42d12","Type":"ContainerDied","Data":"8c2f94b5eacd9b20217771c63cdb899a5ce18dbc2e034a1009a87c031c1625b6"} Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.681459 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c2f94b5eacd9b20217771c63cdb899a5ce18dbc2e034a1009a87c031c1625b6" Oct 10 15:00:03 crc kubenswrapper[4788]: I1010 15:00:03.681186 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn" Oct 10 15:00:08 crc kubenswrapper[4788]: I1010 15:00:08.318369 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5b8b68658d-sn5zq" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.236647 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r"] Oct 10 15:00:09 crc kubenswrapper[4788]: E1010 15:00:09.236979 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3c9bad-9fee-4dd0-b553-52cebeb42d12" containerName="collect-profiles" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.236993 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3c9bad-9fee-4dd0-b553-52cebeb42d12" containerName="collect-profiles" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.237096 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3c9bad-9fee-4dd0-b553-52cebeb42d12" containerName="collect-profiles" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.237602 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.239654 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.239765 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-52mpg" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.245623 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-v9bn2"] Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.251968 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.256662 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r"] Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.259723 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.259742 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.341080 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-8gzh5"] Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.342094 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.343733 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qlmcv" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.343875 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.343880 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.344767 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.368813 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-g8jk6"] Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.370538 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.378975 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.408938 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-reloader\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.408982 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-frr-conf\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.409027 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/85b09555-e996-4aee-8882-5e2bf28f4d1b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-5bj2r\" (UID: \"85b09555-e996-4aee-8882-5e2bf28f4d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.409051 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/22cedde9-131c-445d-832d-8063a6d3adf6-frr-startup\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.409087 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-frr-sockets\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.409104 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d96c\" (UniqueName: \"kubernetes.io/projected/85b09555-e996-4aee-8882-5e2bf28f4d1b-kube-api-access-8d96c\") pod \"frr-k8s-webhook-server-64bf5d555-5bj2r\" (UID: \"85b09555-e996-4aee-8882-5e2bf28f4d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.409131 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-metrics\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.409203 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4fhl\" (UniqueName: \"kubernetes.io/projected/22cedde9-131c-445d-832d-8063a6d3adf6-kube-api-access-w4fhl\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.409222 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22cedde9-131c-445d-832d-8063a6d3adf6-metrics-certs\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.430818 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-g8jk6"] Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510371 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-metrics-certs\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510471 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-frr-sockets\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510533 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d96c\" (UniqueName: \"kubernetes.io/projected/85b09555-e996-4aee-8882-5e2bf28f4d1b-kube-api-access-8d96c\") pod \"frr-k8s-webhook-server-64bf5d555-5bj2r\" (UID: \"85b09555-e996-4aee-8882-5e2bf28f4d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510562 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-metrics\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510589 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jddb4\" (UniqueName: \"kubernetes.io/projected/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-kube-api-access-jddb4\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510625 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4fhl\" (UniqueName: \"kubernetes.io/projected/22cedde9-131c-445d-832d-8063a6d3adf6-kube-api-access-w4fhl\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510654 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22cedde9-131c-445d-832d-8063a6d3adf6-metrics-certs\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510672 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-metrics-certs\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510693 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510711 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-cert\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510731 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-reloader\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510751 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-frr-conf\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510780 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ef3cd7ba-2155-4457-a62c-63660545e0cc-metallb-excludel2\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510800 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/85b09555-e996-4aee-8882-5e2bf28f4d1b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-5bj2r\" (UID: \"85b09555-e996-4aee-8882-5e2bf28f4d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510821 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/22cedde9-131c-445d-832d-8063a6d3adf6-frr-startup\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.510845 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4txbd\" (UniqueName: \"kubernetes.io/projected/ef3cd7ba-2155-4457-a62c-63660545e0cc-kube-api-access-4txbd\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.511108 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-frr-sockets\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.511245 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-metrics\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.511375 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-reloader\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.511433 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/22cedde9-131c-445d-832d-8063a6d3adf6-frr-conf\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.512162 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/22cedde9-131c-445d-832d-8063a6d3adf6-frr-startup\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.518079 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/85b09555-e996-4aee-8882-5e2bf28f4d1b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-5bj2r\" (UID: \"85b09555-e996-4aee-8882-5e2bf28f4d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.534483 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22cedde9-131c-445d-832d-8063a6d3adf6-metrics-certs\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.536906 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d96c\" (UniqueName: \"kubernetes.io/projected/85b09555-e996-4aee-8882-5e2bf28f4d1b-kube-api-access-8d96c\") pod \"frr-k8s-webhook-server-64bf5d555-5bj2r\" (UID: \"85b09555-e996-4aee-8882-5e2bf28f4d1b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.540612 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4fhl\" (UniqueName: \"kubernetes.io/projected/22cedde9-131c-445d-832d-8063a6d3adf6-kube-api-access-w4fhl\") pod \"frr-k8s-v9bn2\" (UID: \"22cedde9-131c-445d-832d-8063a6d3adf6\") " pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.562787 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.570788 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.612159 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-cert\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.612230 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ef3cd7ba-2155-4457-a62c-63660545e0cc-metallb-excludel2\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.612269 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4txbd\" (UniqueName: \"kubernetes.io/projected/ef3cd7ba-2155-4457-a62c-63660545e0cc-kube-api-access-4txbd\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.612290 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-metrics-certs\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.612319 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jddb4\" (UniqueName: \"kubernetes.io/projected/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-kube-api-access-jddb4\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.612366 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-metrics-certs\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.612388 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: E1010 15:00:09.612516 4788 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 10 15:00:09 crc kubenswrapper[4788]: E1010 15:00:09.612577 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist podName:ef3cd7ba-2155-4457-a62c-63660545e0cc nodeName:}" failed. No retries permitted until 2025-10-10 15:00:10.112557749 +0000 UTC m=+912.562273287 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist") pod "speaker-8gzh5" (UID: "ef3cd7ba-2155-4457-a62c-63660545e0cc") : secret "metallb-memberlist" not found Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.613453 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ef3cd7ba-2155-4457-a62c-63660545e0cc-metallb-excludel2\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.614666 4788 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.615923 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-metrics-certs\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.616347 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-metrics-certs\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.624858 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-cert\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.632303 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4txbd\" (UniqueName: \"kubernetes.io/projected/ef3cd7ba-2155-4457-a62c-63660545e0cc-kube-api-access-4txbd\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.633541 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jddb4\" (UniqueName: \"kubernetes.io/projected/02700c1c-08fb-45a3-956b-bdbc6ac4e18a-kube-api-access-jddb4\") pod \"controller-68d546b9d8-g8jk6\" (UID: \"02700c1c-08fb-45a3-956b-bdbc6ac4e18a\") " pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.719804 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.801070 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r"] Oct 10 15:00:09 crc kubenswrapper[4788]: I1010 15:00:09.957128 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-g8jk6"] Oct 10 15:00:09 crc kubenswrapper[4788]: W1010 15:00:09.961984 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02700c1c_08fb_45a3_956b_bdbc6ac4e18a.slice/crio-213012e71202cd58ef2ff74544b893facc19e504f5284f2c7816c54c98d7d475 WatchSource:0}: Error finding container 213012e71202cd58ef2ff74544b893facc19e504f5284f2c7816c54c98d7d475: Status 404 returned error can't find the container with id 213012e71202cd58ef2ff74544b893facc19e504f5284f2c7816c54c98d7d475 Oct 10 15:00:10 crc kubenswrapper[4788]: I1010 15:00:10.120213 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:10 crc kubenswrapper[4788]: E1010 15:00:10.120329 4788 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 10 15:00:10 crc kubenswrapper[4788]: E1010 15:00:10.120375 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist podName:ef3cd7ba-2155-4457-a62c-63660545e0cc nodeName:}" failed. No retries permitted until 2025-10-10 15:00:11.120361027 +0000 UTC m=+913.570076575 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist") pod "speaker-8gzh5" (UID: "ef3cd7ba-2155-4457-a62c-63660545e0cc") : secret "metallb-memberlist" not found Oct 10 15:00:10 crc kubenswrapper[4788]: I1010 15:00:10.716921 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerStarted","Data":"d6a294efcd50318f3fdd31d91ec35baeb51c92168cfcf8b7de365958a28e9385"} Oct 10 15:00:10 crc kubenswrapper[4788]: I1010 15:00:10.717796 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" event={"ID":"85b09555-e996-4aee-8882-5e2bf28f4d1b","Type":"ContainerStarted","Data":"0dd9be78ec7d58d9afd479b338f872dec9aa24e6044e292a0f4d9f554a8a65b2"} Oct 10 15:00:10 crc kubenswrapper[4788]: I1010 15:00:10.719559 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-g8jk6" event={"ID":"02700c1c-08fb-45a3-956b-bdbc6ac4e18a","Type":"ContainerStarted","Data":"f4c62e09a1c41480f08dfba3a8c170878a758515a665a59a08cc7ed7d076730f"} Oct 10 15:00:10 crc kubenswrapper[4788]: I1010 15:00:10.719586 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-g8jk6" event={"ID":"02700c1c-08fb-45a3-956b-bdbc6ac4e18a","Type":"ContainerStarted","Data":"fac2d2aa1500e96a261c15a701e84ac80a0f2c5cdbaa367f82153fe5fb74dcc6"} Oct 10 15:00:10 crc kubenswrapper[4788]: I1010 15:00:10.719603 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-g8jk6" event={"ID":"02700c1c-08fb-45a3-956b-bdbc6ac4e18a","Type":"ContainerStarted","Data":"213012e71202cd58ef2ff74544b893facc19e504f5284f2c7816c54c98d7d475"} Oct 10 15:00:10 crc kubenswrapper[4788]: I1010 15:00:10.719720 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:10 crc kubenswrapper[4788]: I1010 15:00:10.740467 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-g8jk6" podStartSLOduration=1.740447554 podStartE2EDuration="1.740447554s" podCreationTimestamp="2025-10-10 15:00:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:00:10.736561408 +0000 UTC m=+913.186276956" watchObservedRunningTime="2025-10-10 15:00:10.740447554 +0000 UTC m=+913.190163102" Oct 10 15:00:11 crc kubenswrapper[4788]: I1010 15:00:11.133788 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:11 crc kubenswrapper[4788]: I1010 15:00:11.144758 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ef3cd7ba-2155-4457-a62c-63660545e0cc-memberlist\") pod \"speaker-8gzh5\" (UID: \"ef3cd7ba-2155-4457-a62c-63660545e0cc\") " pod="metallb-system/speaker-8gzh5" Oct 10 15:00:11 crc kubenswrapper[4788]: I1010 15:00:11.155776 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8gzh5" Oct 10 15:00:11 crc kubenswrapper[4788]: I1010 15:00:11.736298 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8gzh5" event={"ID":"ef3cd7ba-2155-4457-a62c-63660545e0cc","Type":"ContainerStarted","Data":"e8c14c72c13bc3ccbea9c6fa215f7e3d06d0d4b5edbfcdf047f1097e3e6d1ca9"} Oct 10 15:00:11 crc kubenswrapper[4788]: I1010 15:00:11.736347 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8gzh5" event={"ID":"ef3cd7ba-2155-4457-a62c-63660545e0cc","Type":"ContainerStarted","Data":"50c57a8d31fb7bc11c6776ecdb17a4ca38c27d4b2eed9c7f20b5df7957f3b984"} Oct 10 15:00:12 crc kubenswrapper[4788]: I1010 15:00:12.790723 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8gzh5" event={"ID":"ef3cd7ba-2155-4457-a62c-63660545e0cc","Type":"ContainerStarted","Data":"0845f64b8a91ad36a79379097a7a9a6e7e8d941effd10d0031284ab369ffd034"} Oct 10 15:00:12 crc kubenswrapper[4788]: I1010 15:00:12.793380 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-8gzh5" Oct 10 15:00:12 crc kubenswrapper[4788]: I1010 15:00:12.835753 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-8gzh5" podStartSLOduration=3.835735765 podStartE2EDuration="3.835735765s" podCreationTimestamp="2025-10-10 15:00:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:00:12.832400155 +0000 UTC m=+915.282115703" watchObservedRunningTime="2025-10-10 15:00:12.835735765 +0000 UTC m=+915.285451313" Oct 10 15:00:17 crc kubenswrapper[4788]: I1010 15:00:17.828570 4788 generic.go:334] "Generic (PLEG): container finished" podID="22cedde9-131c-445d-832d-8063a6d3adf6" containerID="abc5a52f783c53a72b955d7c0eb56d988d8319a8dc30729bc5419618527c459f" exitCode=0 Oct 10 15:00:17 crc kubenswrapper[4788]: I1010 15:00:17.828709 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerDied","Data":"abc5a52f783c53a72b955d7c0eb56d988d8319a8dc30729bc5419618527c459f"} Oct 10 15:00:17 crc kubenswrapper[4788]: I1010 15:00:17.830658 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" event={"ID":"85b09555-e996-4aee-8882-5e2bf28f4d1b","Type":"ContainerStarted","Data":"972db5e07f20deace17e0d63ce82c641107bc97196126ad7abbd1f7ec700f522"} Oct 10 15:00:17 crc kubenswrapper[4788]: I1010 15:00:17.830798 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:17 crc kubenswrapper[4788]: I1010 15:00:17.884764 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" podStartSLOduration=1.117962602 podStartE2EDuration="8.884747986s" podCreationTimestamp="2025-10-10 15:00:09 +0000 UTC" firstStartedPulling="2025-10-10 15:00:09.817604637 +0000 UTC m=+912.267320185" lastFinishedPulling="2025-10-10 15:00:17.584390021 +0000 UTC m=+920.034105569" observedRunningTime="2025-10-10 15:00:17.880171762 +0000 UTC m=+920.329887310" watchObservedRunningTime="2025-10-10 15:00:17.884747986 +0000 UTC m=+920.334463524" Oct 10 15:00:18 crc kubenswrapper[4788]: I1010 15:00:18.839055 4788 generic.go:334] "Generic (PLEG): container finished" podID="22cedde9-131c-445d-832d-8063a6d3adf6" containerID="53e580ce18db66d0a3cc9dffd5c46412d293236d45ec00b714167da080be327d" exitCode=0 Oct 10 15:00:18 crc kubenswrapper[4788]: I1010 15:00:18.839171 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerDied","Data":"53e580ce18db66d0a3cc9dffd5c46412d293236d45ec00b714167da080be327d"} Oct 10 15:00:19 crc kubenswrapper[4788]: I1010 15:00:19.847114 4788 generic.go:334] "Generic (PLEG): container finished" podID="22cedde9-131c-445d-832d-8063a6d3adf6" containerID="c6f7273e2e857d0de35e190e098cb3f0d09d6562b026a9b6612785667c484d0d" exitCode=0 Oct 10 15:00:19 crc kubenswrapper[4788]: I1010 15:00:19.847457 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerDied","Data":"c6f7273e2e857d0de35e190e098cb3f0d09d6562b026a9b6612785667c484d0d"} Oct 10 15:00:20 crc kubenswrapper[4788]: I1010 15:00:20.863557 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerStarted","Data":"fef23b37a1029cd78a2daef507d4faf560cb20c391cc87763f3b81dabe693736"} Oct 10 15:00:20 crc kubenswrapper[4788]: I1010 15:00:20.863901 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerStarted","Data":"837cae600df3d595297020704ef7e3a0b9bc0cdca711f16a4f609bbeb8ffe744"} Oct 10 15:00:20 crc kubenswrapper[4788]: I1010 15:00:20.863916 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerStarted","Data":"ea9543efaa1983d88827a9918ea1bff62b658aeb20c006bf9f7489e52d6642e2"} Oct 10 15:00:20 crc kubenswrapper[4788]: I1010 15:00:20.863926 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerStarted","Data":"a41fdd16421d83dae2e9e8e8b64213ecbe561bb3390af4c940bbc7397b04ec9a"} Oct 10 15:00:20 crc kubenswrapper[4788]: I1010 15:00:20.863936 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerStarted","Data":"adabfae483eb34c84c83d956d2199e33a287cfe4d5b519ad4ab6de98ea251f3b"} Oct 10 15:00:21 crc kubenswrapper[4788]: I1010 15:00:21.160532 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-8gzh5" Oct 10 15:00:21 crc kubenswrapper[4788]: I1010 15:00:21.873659 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v9bn2" event={"ID":"22cedde9-131c-445d-832d-8063a6d3adf6","Type":"ContainerStarted","Data":"c2eac55a2569089dc1373db360abb73a071287f700f317084c10c7c3d78f8069"} Oct 10 15:00:21 crc kubenswrapper[4788]: I1010 15:00:21.874700 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:21 crc kubenswrapper[4788]: I1010 15:00:21.903972 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-v9bn2" podStartSLOduration=5.078780526 podStartE2EDuration="12.903943156s" podCreationTimestamp="2025-10-10 15:00:09 +0000 UTC" firstStartedPulling="2025-10-10 15:00:09.754972316 +0000 UTC m=+912.204687874" lastFinishedPulling="2025-10-10 15:00:17.580134956 +0000 UTC m=+920.029850504" observedRunningTime="2025-10-10 15:00:21.901824709 +0000 UTC m=+924.351540287" watchObservedRunningTime="2025-10-10 15:00:21.903943156 +0000 UTC m=+924.353658694" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.039363 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl"] Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.040781 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.042967 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.061122 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl"] Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.067394 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.067470 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.067521 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4f8d\" (UniqueName: \"kubernetes.io/projected/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-kube-api-access-f4f8d\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.168366 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.168412 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.168618 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4f8d\" (UniqueName: \"kubernetes.io/projected/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-kube-api-access-f4f8d\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.169207 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.170118 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.197396 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4f8d\" (UniqueName: \"kubernetes.io/projected/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-kube-api-access-f4f8d\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.359838 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.602381 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl"] Oct 10 15:00:23 crc kubenswrapper[4788]: W1010 15:00:23.611444 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6f44e2f_cadb_4a7c_baa1_2305298df8c5.slice/crio-a35c8cebce104624fd9e44292992f34e7f78c8c8454a86c7c04505f0db9d488f WatchSource:0}: Error finding container a35c8cebce104624fd9e44292992f34e7f78c8c8454a86c7c04505f0db9d488f: Status 404 returned error can't find the container with id a35c8cebce104624fd9e44292992f34e7f78c8c8454a86c7c04505f0db9d488f Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.898736 4788 generic.go:334] "Generic (PLEG): container finished" podID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerID="ad7322a93f0419fb56fb11579e465435c8fb409a39baae16eae7027f71fcc327" exitCode=0 Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.899801 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" event={"ID":"c6f44e2f-cadb-4a7c-baa1-2305298df8c5","Type":"ContainerDied","Data":"ad7322a93f0419fb56fb11579e465435c8fb409a39baae16eae7027f71fcc327"} Oct 10 15:00:23 crc kubenswrapper[4788]: I1010 15:00:23.899861 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" event={"ID":"c6f44e2f-cadb-4a7c-baa1-2305298df8c5","Type":"ContainerStarted","Data":"a35c8cebce104624fd9e44292992f34e7f78c8c8454a86c7c04505f0db9d488f"} Oct 10 15:00:24 crc kubenswrapper[4788]: I1010 15:00:24.571425 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:24 crc kubenswrapper[4788]: I1010 15:00:24.642182 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:27 crc kubenswrapper[4788]: I1010 15:00:27.930012 4788 generic.go:334] "Generic (PLEG): container finished" podID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerID="fc42bec5a392ba328e765b7398909ab7110f6b592788c2f5ab995c6433298172" exitCode=0 Oct 10 15:00:27 crc kubenswrapper[4788]: I1010 15:00:27.930077 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" event={"ID":"c6f44e2f-cadb-4a7c-baa1-2305298df8c5","Type":"ContainerDied","Data":"fc42bec5a392ba328e765b7398909ab7110f6b592788c2f5ab995c6433298172"} Oct 10 15:00:28 crc kubenswrapper[4788]: I1010 15:00:28.941508 4788 generic.go:334] "Generic (PLEG): container finished" podID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerID="bb01c983ca9c6020497bf84e1a3237b819e0792500651a7dd2d29bd3cf0f89df" exitCode=0 Oct 10 15:00:28 crc kubenswrapper[4788]: I1010 15:00:28.941580 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" event={"ID":"c6f44e2f-cadb-4a7c-baa1-2305298df8c5","Type":"ContainerDied","Data":"bb01c983ca9c6020497bf84e1a3237b819e0792500651a7dd2d29bd3cf0f89df"} Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.405936 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.406003 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.406052 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.406752 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d62f45676977d2854ac3c581616b717de6d75ef758d15d869e78e7d7a3a145f9"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.406825 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://d62f45676977d2854ac3c581616b717de6d75ef758d15d869e78e7d7a3a145f9" gracePeriod=600 Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.569271 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-5bj2r" Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.723249 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-g8jk6" Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.950087 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="d62f45676977d2854ac3c581616b717de6d75ef758d15d869e78e7d7a3a145f9" exitCode=0 Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.950190 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"d62f45676977d2854ac3c581616b717de6d75ef758d15d869e78e7d7a3a145f9"} Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.950244 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"9bf2bdef66313687037ed74946bc83398b51a4a5e0533a877fae35897091d486"} Oct 10 15:00:29 crc kubenswrapper[4788]: I1010 15:00:29.950280 4788 scope.go:117] "RemoveContainer" containerID="10d046d777ded5357953f35d670fff38bbf3941e6d3d9f85f546d304830131b6" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.319510 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.512603 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4f8d\" (UniqueName: \"kubernetes.io/projected/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-kube-api-access-f4f8d\") pod \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.512797 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-bundle\") pod \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.512989 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-util\") pod \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\" (UID: \"c6f44e2f-cadb-4a7c-baa1-2305298df8c5\") " Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.514246 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-bundle" (OuterVolumeSpecName: "bundle") pod "c6f44e2f-cadb-4a7c-baa1-2305298df8c5" (UID: "c6f44e2f-cadb-4a7c-baa1-2305298df8c5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.522923 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-util" (OuterVolumeSpecName: "util") pod "c6f44e2f-cadb-4a7c-baa1-2305298df8c5" (UID: "c6f44e2f-cadb-4a7c-baa1-2305298df8c5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.534821 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-kube-api-access-f4f8d" (OuterVolumeSpecName: "kube-api-access-f4f8d") pod "c6f44e2f-cadb-4a7c-baa1-2305298df8c5" (UID: "c6f44e2f-cadb-4a7c-baa1-2305298df8c5"). InnerVolumeSpecName "kube-api-access-f4f8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.614974 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4f8d\" (UniqueName: \"kubernetes.io/projected/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-kube-api-access-f4f8d\") on node \"crc\" DevicePath \"\"" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.615018 4788 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.615030 4788 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6f44e2f-cadb-4a7c-baa1-2305298df8c5-util\") on node \"crc\" DevicePath \"\"" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.964882 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" event={"ID":"c6f44e2f-cadb-4a7c-baa1-2305298df8c5","Type":"ContainerDied","Data":"a35c8cebce104624fd9e44292992f34e7f78c8c8454a86c7c04505f0db9d488f"} Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.964929 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a35c8cebce104624fd9e44292992f34e7f78c8c8454a86c7c04505f0db9d488f" Oct 10 15:00:30 crc kubenswrapper[4788]: I1010 15:00:30.964968 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.325597 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn"] Oct 10 15:00:38 crc kubenswrapper[4788]: E1010 15:00:38.326600 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerName="pull" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.326616 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerName="pull" Oct 10 15:00:38 crc kubenswrapper[4788]: E1010 15:00:38.326643 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerName="util" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.326652 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerName="util" Oct 10 15:00:38 crc kubenswrapper[4788]: E1010 15:00:38.326663 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerName="extract" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.326670 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerName="extract" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.326795 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6f44e2f-cadb-4a7c-baa1-2305298df8c5" containerName="extract" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.327293 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.329775 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.331251 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.344014 4788 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-glw8j" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.346740 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8czxz\" (UniqueName: \"kubernetes.io/projected/87f3960d-b04f-47c1-bf1d-2ffd533b5e87-kube-api-access-8czxz\") pod \"cert-manager-operator-controller-manager-57cd46d6d-9rjkn\" (UID: \"87f3960d-b04f-47c1-bf1d-2ffd533b5e87\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.367580 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn"] Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.448209 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8czxz\" (UniqueName: \"kubernetes.io/projected/87f3960d-b04f-47c1-bf1d-2ffd533b5e87-kube-api-access-8czxz\") pod \"cert-manager-operator-controller-manager-57cd46d6d-9rjkn\" (UID: \"87f3960d-b04f-47c1-bf1d-2ffd533b5e87\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.477744 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8czxz\" (UniqueName: \"kubernetes.io/projected/87f3960d-b04f-47c1-bf1d-2ffd533b5e87-kube-api-access-8czxz\") pod \"cert-manager-operator-controller-manager-57cd46d6d-9rjkn\" (UID: \"87f3960d-b04f-47c1-bf1d-2ffd533b5e87\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn" Oct 10 15:00:38 crc kubenswrapper[4788]: I1010 15:00:38.670611 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn" Oct 10 15:00:39 crc kubenswrapper[4788]: I1010 15:00:39.057557 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn"] Oct 10 15:00:39 crc kubenswrapper[4788]: W1010 15:00:39.064695 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87f3960d_b04f_47c1_bf1d_2ffd533b5e87.slice/crio-a870f8c3e2065f262d09f80a4edb3f6b16ba53b9ada8e920bbbbfe1abeb110bf WatchSource:0}: Error finding container a870f8c3e2065f262d09f80a4edb3f6b16ba53b9ada8e920bbbbfe1abeb110bf: Status 404 returned error can't find the container with id a870f8c3e2065f262d09f80a4edb3f6b16ba53b9ada8e920bbbbfe1abeb110bf Oct 10 15:00:39 crc kubenswrapper[4788]: I1010 15:00:39.574271 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-v9bn2" Oct 10 15:00:40 crc kubenswrapper[4788]: I1010 15:00:40.032894 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn" event={"ID":"87f3960d-b04f-47c1-bf1d-2ffd533b5e87","Type":"ContainerStarted","Data":"a870f8c3e2065f262d09f80a4edb3f6b16ba53b9ada8e920bbbbfe1abeb110bf"} Oct 10 15:00:48 crc kubenswrapper[4788]: I1010 15:00:48.096374 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn" event={"ID":"87f3960d-b04f-47c1-bf1d-2ffd533b5e87","Type":"ContainerStarted","Data":"d3aaa2de1e9a13440f275a593e54fb09a3a8df701ee2c38ba1f479d4f072997b"} Oct 10 15:00:48 crc kubenswrapper[4788]: I1010 15:00:48.127458 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-9rjkn" podStartSLOduration=2.047325166 podStartE2EDuration="10.127440488s" podCreationTimestamp="2025-10-10 15:00:38 +0000 UTC" firstStartedPulling="2025-10-10 15:00:39.06890438 +0000 UTC m=+941.518619938" lastFinishedPulling="2025-10-10 15:00:47.149019712 +0000 UTC m=+949.598735260" observedRunningTime="2025-10-10 15:00:48.123557562 +0000 UTC m=+950.573273300" watchObservedRunningTime="2025-10-10 15:00:48.127440488 +0000 UTC m=+950.577156036" Oct 10 15:00:51 crc kubenswrapper[4788]: I1010 15:00:51.959047 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-5jvst"] Oct 10 15:00:51 crc kubenswrapper[4788]: I1010 15:00:51.960724 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:51 crc kubenswrapper[4788]: I1010 15:00:51.962721 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 10 15:00:51 crc kubenswrapper[4788]: I1010 15:00:51.962782 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 10 15:00:51 crc kubenswrapper[4788]: I1010 15:00:51.964123 4788 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-2sw8s" Oct 10 15:00:51 crc kubenswrapper[4788]: I1010 15:00:51.974238 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-5jvst"] Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.075587 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78vpk\" (UniqueName: \"kubernetes.io/projected/921d9dbb-d20b-4a32-b7ad-ff9ab5e694da-kube-api-access-78vpk\") pod \"cert-manager-webhook-d969966f-5jvst\" (UID: \"921d9dbb-d20b-4a32-b7ad-ff9ab5e694da\") " pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.075675 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/921d9dbb-d20b-4a32-b7ad-ff9ab5e694da-bound-sa-token\") pod \"cert-manager-webhook-d969966f-5jvst\" (UID: \"921d9dbb-d20b-4a32-b7ad-ff9ab5e694da\") " pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.178198 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/921d9dbb-d20b-4a32-b7ad-ff9ab5e694da-bound-sa-token\") pod \"cert-manager-webhook-d969966f-5jvst\" (UID: \"921d9dbb-d20b-4a32-b7ad-ff9ab5e694da\") " pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.178324 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78vpk\" (UniqueName: \"kubernetes.io/projected/921d9dbb-d20b-4a32-b7ad-ff9ab5e694da-kube-api-access-78vpk\") pod \"cert-manager-webhook-d969966f-5jvst\" (UID: \"921d9dbb-d20b-4a32-b7ad-ff9ab5e694da\") " pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.204597 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/921d9dbb-d20b-4a32-b7ad-ff9ab5e694da-bound-sa-token\") pod \"cert-manager-webhook-d969966f-5jvst\" (UID: \"921d9dbb-d20b-4a32-b7ad-ff9ab5e694da\") " pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.204654 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78vpk\" (UniqueName: \"kubernetes.io/projected/921d9dbb-d20b-4a32-b7ad-ff9ab5e694da-kube-api-access-78vpk\") pod \"cert-manager-webhook-d969966f-5jvst\" (UID: \"921d9dbb-d20b-4a32-b7ad-ff9ab5e694da\") " pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.281403 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.529072 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-5jvst"] Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.645840 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n"] Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.646822 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.651901 4788 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-fb2fr" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.658515 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n"] Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.786064 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpmgr\" (UniqueName: \"kubernetes.io/projected/8c24f8c3-1871-44fb-a84c-e4d5b5885604-kube-api-access-lpmgr\") pod \"cert-manager-cainjector-7d9f95dbf-mgv2n\" (UID: \"8c24f8c3-1871-44fb-a84c-e4d5b5885604\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.786270 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c24f8c3-1871-44fb-a84c-e4d5b5885604-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-mgv2n\" (UID: \"8c24f8c3-1871-44fb-a84c-e4d5b5885604\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.887076 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpmgr\" (UniqueName: \"kubernetes.io/projected/8c24f8c3-1871-44fb-a84c-e4d5b5885604-kube-api-access-lpmgr\") pod \"cert-manager-cainjector-7d9f95dbf-mgv2n\" (UID: \"8c24f8c3-1871-44fb-a84c-e4d5b5885604\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.887192 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c24f8c3-1871-44fb-a84c-e4d5b5885604-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-mgv2n\" (UID: \"8c24f8c3-1871-44fb-a84c-e4d5b5885604\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.905192 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c24f8c3-1871-44fb-a84c-e4d5b5885604-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-mgv2n\" (UID: \"8c24f8c3-1871-44fb-a84c-e4d5b5885604\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.905546 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpmgr\" (UniqueName: \"kubernetes.io/projected/8c24f8c3-1871-44fb-a84c-e4d5b5885604-kube-api-access-lpmgr\") pod \"cert-manager-cainjector-7d9f95dbf-mgv2n\" (UID: \"8c24f8c3-1871-44fb-a84c-e4d5b5885604\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" Oct 10 15:00:52 crc kubenswrapper[4788]: I1010 15:00:52.971752 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" Oct 10 15:00:53 crc kubenswrapper[4788]: I1010 15:00:53.132415 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-5jvst" event={"ID":"921d9dbb-d20b-4a32-b7ad-ff9ab5e694da","Type":"ContainerStarted","Data":"b073fbd051f2f5c49e7186edb87742cffdf0cf7efb47624b97bb2eb42db233ff"} Oct 10 15:00:53 crc kubenswrapper[4788]: I1010 15:00:53.400130 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n"] Oct 10 15:00:53 crc kubenswrapper[4788]: W1010 15:00:53.404753 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c24f8c3_1871_44fb_a84c_e4d5b5885604.slice/crio-101bd6d42c0b1ad39ff521ab3bddc71a603333d3d2f13168df4627b220a6a5f0 WatchSource:0}: Error finding container 101bd6d42c0b1ad39ff521ab3bddc71a603333d3d2f13168df4627b220a6a5f0: Status 404 returned error can't find the container with id 101bd6d42c0b1ad39ff521ab3bddc71a603333d3d2f13168df4627b220a6a5f0 Oct 10 15:00:54 crc kubenswrapper[4788]: I1010 15:00:54.143503 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" event={"ID":"8c24f8c3-1871-44fb-a84c-e4d5b5885604","Type":"ContainerStarted","Data":"101bd6d42c0b1ad39ff521ab3bddc71a603333d3d2f13168df4627b220a6a5f0"} Oct 10 15:00:58 crc kubenswrapper[4788]: I1010 15:00:58.169482 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-5jvst" event={"ID":"921d9dbb-d20b-4a32-b7ad-ff9ab5e694da","Type":"ContainerStarted","Data":"bd6b282813576b7529e19670b74f1517c18bfb0ad60d5850ef16a520f3879f7a"} Oct 10 15:00:58 crc kubenswrapper[4788]: I1010 15:00:58.170382 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:00:58 crc kubenswrapper[4788]: I1010 15:00:58.171564 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" event={"ID":"8c24f8c3-1871-44fb-a84c-e4d5b5885604","Type":"ContainerStarted","Data":"e4f560d040820bc0bbcf256f6cad70d59728f8ea8601e71f03b42fce502e581a"} Oct 10 15:00:58 crc kubenswrapper[4788]: I1010 15:00:58.211308 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-5jvst" podStartSLOduration=2.669326672 podStartE2EDuration="7.211292706s" podCreationTimestamp="2025-10-10 15:00:51 +0000 UTC" firstStartedPulling="2025-10-10 15:00:52.547280986 +0000 UTC m=+954.996996534" lastFinishedPulling="2025-10-10 15:00:57.08924702 +0000 UTC m=+959.538962568" observedRunningTime="2025-10-10 15:00:58.20959576 +0000 UTC m=+960.659311308" watchObservedRunningTime="2025-10-10 15:00:58.211292706 +0000 UTC m=+960.661008254" Oct 10 15:00:58 crc kubenswrapper[4788]: I1010 15:00:58.226389 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-mgv2n" podStartSLOduration=2.524513772 podStartE2EDuration="6.226371586s" podCreationTimestamp="2025-10-10 15:00:52 +0000 UTC" firstStartedPulling="2025-10-10 15:00:53.407853292 +0000 UTC m=+955.857568860" lastFinishedPulling="2025-10-10 15:00:57.109711126 +0000 UTC m=+959.559426674" observedRunningTime="2025-10-10 15:00:58.224871655 +0000 UTC m=+960.674587203" watchObservedRunningTime="2025-10-10 15:00:58.226371586 +0000 UTC m=+960.676087134" Oct 10 15:01:02 crc kubenswrapper[4788]: I1010 15:01:02.284750 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-5jvst" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.204668 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wgz9w"] Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.205808 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.232008 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgz9w"] Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.385605 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxkkh\" (UniqueName: \"kubernetes.io/projected/b0fb857d-035d-43c7-bb42-4f523b363abd-kube-api-access-qxkkh\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.385661 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-catalog-content\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.385699 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-utilities\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.487494 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxkkh\" (UniqueName: \"kubernetes.io/projected/b0fb857d-035d-43c7-bb42-4f523b363abd-kube-api-access-qxkkh\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.487556 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-catalog-content\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.487605 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-utilities\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.488157 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-catalog-content\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.488522 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-utilities\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.510726 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxkkh\" (UniqueName: \"kubernetes.io/projected/b0fb857d-035d-43c7-bb42-4f523b363abd-kube-api-access-qxkkh\") pod \"redhat-marketplace-wgz9w\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.527100 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:05 crc kubenswrapper[4788]: I1010 15:01:05.769844 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgz9w"] Oct 10 15:01:06 crc kubenswrapper[4788]: I1010 15:01:06.221825 4788 generic.go:334] "Generic (PLEG): container finished" podID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerID="d12aaff7715b4a7098e581607c52195a3935b94a41dab97739e189320123db3e" exitCode=0 Oct 10 15:01:06 crc kubenswrapper[4788]: I1010 15:01:06.223246 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgz9w" event={"ID":"b0fb857d-035d-43c7-bb42-4f523b363abd","Type":"ContainerDied","Data":"d12aaff7715b4a7098e581607c52195a3935b94a41dab97739e189320123db3e"} Oct 10 15:01:06 crc kubenswrapper[4788]: I1010 15:01:06.223339 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgz9w" event={"ID":"b0fb857d-035d-43c7-bb42-4f523b363abd","Type":"ContainerStarted","Data":"ecfeace83b83d17f17989df9ed7b1db0efb2515796f2560633d59594d9f9e787"} Oct 10 15:01:08 crc kubenswrapper[4788]: I1010 15:01:08.244283 4788 generic.go:334] "Generic (PLEG): container finished" podID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerID="8c527fda5df4ee447930092effda97878464c2e8d983745efc38b3b7034091bc" exitCode=0 Oct 10 15:01:08 crc kubenswrapper[4788]: I1010 15:01:08.252122 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgz9w" event={"ID":"b0fb857d-035d-43c7-bb42-4f523b363abd","Type":"ContainerDied","Data":"8c527fda5df4ee447930092effda97878464c2e8d983745efc38b3b7034091bc"} Oct 10 15:01:09 crc kubenswrapper[4788]: I1010 15:01:09.254615 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgz9w" event={"ID":"b0fb857d-035d-43c7-bb42-4f523b363abd","Type":"ContainerStarted","Data":"30c02c422f342981bc87cbbe4d33586ace65313d1773ac9ef406591887724bd1"} Oct 10 15:01:09 crc kubenswrapper[4788]: I1010 15:01:09.274645 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wgz9w" podStartSLOduration=1.749163567 podStartE2EDuration="4.274565656s" podCreationTimestamp="2025-10-10 15:01:05 +0000 UTC" firstStartedPulling="2025-10-10 15:01:06.224921582 +0000 UTC m=+968.674637130" lastFinishedPulling="2025-10-10 15:01:08.750323661 +0000 UTC m=+971.200039219" observedRunningTime="2025-10-10 15:01:09.270715551 +0000 UTC m=+971.720431099" watchObservedRunningTime="2025-10-10 15:01:09.274565656 +0000 UTC m=+971.724281224" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.598340 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-lhrlm"] Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.599565 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.602900 4788 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-swj2s" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.611061 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-lhrlm"] Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.777155 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6nmd\" (UniqueName: \"kubernetes.io/projected/1ab62119-f153-45ab-8b1a-073493aa5f4b-kube-api-access-t6nmd\") pod \"cert-manager-7d4cc89fcb-lhrlm\" (UID: \"1ab62119-f153-45ab-8b1a-073493aa5f4b\") " pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.777633 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ab62119-f153-45ab-8b1a-073493aa5f4b-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-lhrlm\" (UID: \"1ab62119-f153-45ab-8b1a-073493aa5f4b\") " pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.878907 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6nmd\" (UniqueName: \"kubernetes.io/projected/1ab62119-f153-45ab-8b1a-073493aa5f4b-kube-api-access-t6nmd\") pod \"cert-manager-7d4cc89fcb-lhrlm\" (UID: \"1ab62119-f153-45ab-8b1a-073493aa5f4b\") " pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.878981 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ab62119-f153-45ab-8b1a-073493aa5f4b-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-lhrlm\" (UID: \"1ab62119-f153-45ab-8b1a-073493aa5f4b\") " pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.901069 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6nmd\" (UniqueName: \"kubernetes.io/projected/1ab62119-f153-45ab-8b1a-073493aa5f4b-kube-api-access-t6nmd\") pod \"cert-manager-7d4cc89fcb-lhrlm\" (UID: \"1ab62119-f153-45ab-8b1a-073493aa5f4b\") " pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.904986 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ab62119-f153-45ab-8b1a-073493aa5f4b-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-lhrlm\" (UID: \"1ab62119-f153-45ab-8b1a-073493aa5f4b\") " pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" Oct 10 15:01:11 crc kubenswrapper[4788]: I1010 15:01:11.922742 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" Oct 10 15:01:12 crc kubenswrapper[4788]: I1010 15:01:12.186963 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-lhrlm"] Oct 10 15:01:12 crc kubenswrapper[4788]: W1010 15:01:12.190460 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ab62119_f153_45ab_8b1a_073493aa5f4b.slice/crio-dbc8ba16f6a42dbf15590933493a9f439e79cef68ed18a69521f67d7338a7cbf WatchSource:0}: Error finding container dbc8ba16f6a42dbf15590933493a9f439e79cef68ed18a69521f67d7338a7cbf: Status 404 returned error can't find the container with id dbc8ba16f6a42dbf15590933493a9f439e79cef68ed18a69521f67d7338a7cbf Oct 10 15:01:12 crc kubenswrapper[4788]: I1010 15:01:12.293299 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" event={"ID":"1ab62119-f153-45ab-8b1a-073493aa5f4b","Type":"ContainerStarted","Data":"dbc8ba16f6a42dbf15590933493a9f439e79cef68ed18a69521f67d7338a7cbf"} Oct 10 15:01:13 crc kubenswrapper[4788]: I1010 15:01:13.303849 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" event={"ID":"1ab62119-f153-45ab-8b1a-073493aa5f4b","Type":"ContainerStarted","Data":"66d646ef9ce625dc521c507e6f11d2bbf5801e685dcc846ef08533f0ca47a74e"} Oct 10 15:01:13 crc kubenswrapper[4788]: I1010 15:01:13.330790 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-lhrlm" podStartSLOduration=2.330774848 podStartE2EDuration="2.330774848s" podCreationTimestamp="2025-10-10 15:01:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:01:13.328281341 +0000 UTC m=+975.777996889" watchObservedRunningTime="2025-10-10 15:01:13.330774848 +0000 UTC m=+975.780490396" Oct 10 15:01:15 crc kubenswrapper[4788]: I1010 15:01:15.528504 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:15 crc kubenswrapper[4788]: I1010 15:01:15.529092 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:15 crc kubenswrapper[4788]: I1010 15:01:15.589739 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:16 crc kubenswrapper[4788]: I1010 15:01:16.391725 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:19 crc kubenswrapper[4788]: I1010 15:01:19.185493 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgz9w"] Oct 10 15:01:19 crc kubenswrapper[4788]: I1010 15:01:19.186101 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wgz9w" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerName="registry-server" containerID="cri-o://30c02c422f342981bc87cbbe4d33586ace65313d1773ac9ef406591887724bd1" gracePeriod=2 Oct 10 15:01:19 crc kubenswrapper[4788]: I1010 15:01:19.354971 4788 generic.go:334] "Generic (PLEG): container finished" podID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerID="30c02c422f342981bc87cbbe4d33586ace65313d1773ac9ef406591887724bd1" exitCode=0 Oct 10 15:01:19 crc kubenswrapper[4788]: I1010 15:01:19.355030 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgz9w" event={"ID":"b0fb857d-035d-43c7-bb42-4f523b363abd","Type":"ContainerDied","Data":"30c02c422f342981bc87cbbe4d33586ace65313d1773ac9ef406591887724bd1"} Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.118559 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.147013 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-catalog-content\") pod \"b0fb857d-035d-43c7-bb42-4f523b363abd\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.147233 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-utilities\") pod \"b0fb857d-035d-43c7-bb42-4f523b363abd\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.147289 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxkkh\" (UniqueName: \"kubernetes.io/projected/b0fb857d-035d-43c7-bb42-4f523b363abd-kube-api-access-qxkkh\") pod \"b0fb857d-035d-43c7-bb42-4f523b363abd\" (UID: \"b0fb857d-035d-43c7-bb42-4f523b363abd\") " Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.150558 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-utilities" (OuterVolumeSpecName: "utilities") pod "b0fb857d-035d-43c7-bb42-4f523b363abd" (UID: "b0fb857d-035d-43c7-bb42-4f523b363abd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.160100 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0fb857d-035d-43c7-bb42-4f523b363abd" (UID: "b0fb857d-035d-43c7-bb42-4f523b363abd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.190421 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0fb857d-035d-43c7-bb42-4f523b363abd-kube-api-access-qxkkh" (OuterVolumeSpecName: "kube-api-access-qxkkh") pod "b0fb857d-035d-43c7-bb42-4f523b363abd" (UID: "b0fb857d-035d-43c7-bb42-4f523b363abd"). InnerVolumeSpecName "kube-api-access-qxkkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.250005 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.250577 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0fb857d-035d-43c7-bb42-4f523b363abd-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.250594 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxkkh\" (UniqueName: \"kubernetes.io/projected/b0fb857d-035d-43c7-bb42-4f523b363abd-kube-api-access-qxkkh\") on node \"crc\" DevicePath \"\"" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.368636 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wgz9w" event={"ID":"b0fb857d-035d-43c7-bb42-4f523b363abd","Type":"ContainerDied","Data":"ecfeace83b83d17f17989df9ed7b1db0efb2515796f2560633d59594d9f9e787"} Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.368745 4788 scope.go:117] "RemoveContainer" containerID="30c02c422f342981bc87cbbe4d33586ace65313d1773ac9ef406591887724bd1" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.368929 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wgz9w" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.392892 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgz9w"] Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.396905 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wgz9w"] Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.411757 4788 scope.go:117] "RemoveContainer" containerID="8c527fda5df4ee447930092effda97878464c2e8d983745efc38b3b7034091bc" Oct 10 15:01:20 crc kubenswrapper[4788]: I1010 15:01:20.430030 4788 scope.go:117] "RemoveContainer" containerID="d12aaff7715b4a7098e581607c52195a3935b94a41dab97739e189320123db3e" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.587056 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-68c4c"] Oct 10 15:01:21 crc kubenswrapper[4788]: E1010 15:01:21.587916 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerName="extract-utilities" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.587935 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerName="extract-utilities" Oct 10 15:01:21 crc kubenswrapper[4788]: E1010 15:01:21.587976 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerName="registry-server" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.587985 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerName="registry-server" Oct 10 15:01:21 crc kubenswrapper[4788]: E1010 15:01:21.587994 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerName="extract-content" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.588003 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerName="extract-content" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.588130 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" containerName="registry-server" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.588936 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-68c4c" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.592690 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-8kzph" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.592902 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.594207 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.599598 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-68c4c"] Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.669543 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwv4f\" (UniqueName: \"kubernetes.io/projected/68c2a369-dffb-4791-916f-d74de2c7e292-kube-api-access-nwv4f\") pod \"openstack-operator-index-68c4c\" (UID: \"68c2a369-dffb-4791-916f-d74de2c7e292\") " pod="openstack-operators/openstack-operator-index-68c4c" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.770743 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwv4f\" (UniqueName: \"kubernetes.io/projected/68c2a369-dffb-4791-916f-d74de2c7e292-kube-api-access-nwv4f\") pod \"openstack-operator-index-68c4c\" (UID: \"68c2a369-dffb-4791-916f-d74de2c7e292\") " pod="openstack-operators/openstack-operator-index-68c4c" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.788624 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwv4f\" (UniqueName: \"kubernetes.io/projected/68c2a369-dffb-4791-916f-d74de2c7e292-kube-api-access-nwv4f\") pod \"openstack-operator-index-68c4c\" (UID: \"68c2a369-dffb-4791-916f-d74de2c7e292\") " pod="openstack-operators/openstack-operator-index-68c4c" Oct 10 15:01:21 crc kubenswrapper[4788]: I1010 15:01:21.915697 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-68c4c" Oct 10 15:01:22 crc kubenswrapper[4788]: I1010 15:01:22.240581 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0fb857d-035d-43c7-bb42-4f523b363abd" path="/var/lib/kubelet/pods/b0fb857d-035d-43c7-bb42-4f523b363abd/volumes" Oct 10 15:01:22 crc kubenswrapper[4788]: I1010 15:01:22.303805 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-68c4c"] Oct 10 15:01:22 crc kubenswrapper[4788]: I1010 15:01:22.388936 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-68c4c" event={"ID":"68c2a369-dffb-4791-916f-d74de2c7e292","Type":"ContainerStarted","Data":"d13b22c3b98265653d9d21d3de475a94ab9cee578c5ca3057231210e38b7873b"} Oct 10 15:01:25 crc kubenswrapper[4788]: I1010 15:01:25.412689 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-68c4c" event={"ID":"68c2a369-dffb-4791-916f-d74de2c7e292","Type":"ContainerStarted","Data":"6e2600e466c5732ccabbf5060f6889fc4d3fa66fdbb7dae13162e88390dc24e5"} Oct 10 15:01:25 crc kubenswrapper[4788]: I1010 15:01:25.436774 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-68c4c" podStartSLOduration=2.448989248 podStartE2EDuration="4.436752651s" podCreationTimestamp="2025-10-10 15:01:21 +0000 UTC" firstStartedPulling="2025-10-10 15:01:22.319836877 +0000 UTC m=+984.769552425" lastFinishedPulling="2025-10-10 15:01:24.30760028 +0000 UTC m=+986.757315828" observedRunningTime="2025-10-10 15:01:25.430721938 +0000 UTC m=+987.880437496" watchObservedRunningTime="2025-10-10 15:01:25.436752651 +0000 UTC m=+987.886468219" Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.183346 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-68c4c"] Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.183582 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-68c4c" podUID="68c2a369-dffb-4791-916f-d74de2c7e292" containerName="registry-server" containerID="cri-o://6e2600e466c5732ccabbf5060f6889fc4d3fa66fdbb7dae13162e88390dc24e5" gracePeriod=2 Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.434087 4788 generic.go:334] "Generic (PLEG): container finished" podID="68c2a369-dffb-4791-916f-d74de2c7e292" containerID="6e2600e466c5732ccabbf5060f6889fc4d3fa66fdbb7dae13162e88390dc24e5" exitCode=0 Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.434198 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-68c4c" event={"ID":"68c2a369-dffb-4791-916f-d74de2c7e292","Type":"ContainerDied","Data":"6e2600e466c5732ccabbf5060f6889fc4d3fa66fdbb7dae13162e88390dc24e5"} Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.592301 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-68c4c" Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.724726 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwv4f\" (UniqueName: \"kubernetes.io/projected/68c2a369-dffb-4791-916f-d74de2c7e292-kube-api-access-nwv4f\") pod \"68c2a369-dffb-4791-916f-d74de2c7e292\" (UID: \"68c2a369-dffb-4791-916f-d74de2c7e292\") " Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.734485 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68c2a369-dffb-4791-916f-d74de2c7e292-kube-api-access-nwv4f" (OuterVolumeSpecName: "kube-api-access-nwv4f") pod "68c2a369-dffb-4791-916f-d74de2c7e292" (UID: "68c2a369-dffb-4791-916f-d74de2c7e292"). InnerVolumeSpecName "kube-api-access-nwv4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.826915 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwv4f\" (UniqueName: \"kubernetes.io/projected/68c2a369-dffb-4791-916f-d74de2c7e292-kube-api-access-nwv4f\") on node \"crc\" DevicePath \"\"" Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.983515 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mxkqp"] Oct 10 15:01:28 crc kubenswrapper[4788]: E1010 15:01:28.983840 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68c2a369-dffb-4791-916f-d74de2c7e292" containerName="registry-server" Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.983855 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="68c2a369-dffb-4791-916f-d74de2c7e292" containerName="registry-server" Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.983998 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="68c2a369-dffb-4791-916f-d74de2c7e292" containerName="registry-server" Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.984517 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:28 crc kubenswrapper[4788]: I1010 15:01:28.992597 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mxkqp"] Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.131724 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv8xn\" (UniqueName: \"kubernetes.io/projected/1a2d26de-5eca-4761-b78e-c259524b68f8-kube-api-access-tv8xn\") pod \"openstack-operator-index-mxkqp\" (UID: \"1a2d26de-5eca-4761-b78e-c259524b68f8\") " pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.233980 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv8xn\" (UniqueName: \"kubernetes.io/projected/1a2d26de-5eca-4761-b78e-c259524b68f8-kube-api-access-tv8xn\") pod \"openstack-operator-index-mxkqp\" (UID: \"1a2d26de-5eca-4761-b78e-c259524b68f8\") " pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.253749 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv8xn\" (UniqueName: \"kubernetes.io/projected/1a2d26de-5eca-4761-b78e-c259524b68f8-kube-api-access-tv8xn\") pod \"openstack-operator-index-mxkqp\" (UID: \"1a2d26de-5eca-4761-b78e-c259524b68f8\") " pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.298913 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.449297 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-68c4c" event={"ID":"68c2a369-dffb-4791-916f-d74de2c7e292","Type":"ContainerDied","Data":"d13b22c3b98265653d9d21d3de475a94ab9cee578c5ca3057231210e38b7873b"} Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.449361 4788 scope.go:117] "RemoveContainer" containerID="6e2600e466c5732ccabbf5060f6889fc4d3fa66fdbb7dae13162e88390dc24e5" Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.449493 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-68c4c" Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.491957 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-68c4c"] Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.495723 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-68c4c"] Oct 10 15:01:29 crc kubenswrapper[4788]: I1010 15:01:29.735513 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mxkqp"] Oct 10 15:01:29 crc kubenswrapper[4788]: W1010 15:01:29.745720 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a2d26de_5eca_4761_b78e_c259524b68f8.slice/crio-e66f8b6e2e21a6d13382db0d0950516f78c8e2f5c479ab0ebf95dc765241b126 WatchSource:0}: Error finding container e66f8b6e2e21a6d13382db0d0950516f78c8e2f5c479ab0ebf95dc765241b126: Status 404 returned error can't find the container with id e66f8b6e2e21a6d13382db0d0950516f78c8e2f5c479ab0ebf95dc765241b126 Oct 10 15:01:30 crc kubenswrapper[4788]: I1010 15:01:30.247641 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68c2a369-dffb-4791-916f-d74de2c7e292" path="/var/lib/kubelet/pods/68c2a369-dffb-4791-916f-d74de2c7e292/volumes" Oct 10 15:01:30 crc kubenswrapper[4788]: I1010 15:01:30.455705 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mxkqp" event={"ID":"1a2d26de-5eca-4761-b78e-c259524b68f8","Type":"ContainerStarted","Data":"b9a109707cd90d612633d8b722ee641cd801976a2000fee5b4f1f0e28cde38ef"} Oct 10 15:01:30 crc kubenswrapper[4788]: I1010 15:01:30.455755 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mxkqp" event={"ID":"1a2d26de-5eca-4761-b78e-c259524b68f8","Type":"ContainerStarted","Data":"e66f8b6e2e21a6d13382db0d0950516f78c8e2f5c479ab0ebf95dc765241b126"} Oct 10 15:01:30 crc kubenswrapper[4788]: I1010 15:01:30.476937 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mxkqp" podStartSLOduration=2.432315787 podStartE2EDuration="2.476912484s" podCreationTimestamp="2025-10-10 15:01:28 +0000 UTC" firstStartedPulling="2025-10-10 15:01:29.748408219 +0000 UTC m=+992.198123777" lastFinishedPulling="2025-10-10 15:01:29.793004916 +0000 UTC m=+992.242720474" observedRunningTime="2025-10-10 15:01:30.471447136 +0000 UTC m=+992.921162724" watchObservedRunningTime="2025-10-10 15:01:30.476912484 +0000 UTC m=+992.926628022" Oct 10 15:01:39 crc kubenswrapper[4788]: I1010 15:01:39.299958 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:39 crc kubenswrapper[4788]: I1010 15:01:39.301183 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:39 crc kubenswrapper[4788]: I1010 15:01:39.346774 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:39 crc kubenswrapper[4788]: I1010 15:01:39.564510 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-mxkqp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.253341 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp"] Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.256529 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.259940 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-chd4v" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.267578 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp"] Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.350238 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-util\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.350465 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5c7k\" (UniqueName: \"kubernetes.io/projected/a7f6140b-914c-4de9-9ab3-11cb6fd83225-kube-api-access-q5c7k\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.350607 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-bundle\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.452230 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-util\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.452313 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5c7k\" (UniqueName: \"kubernetes.io/projected/a7f6140b-914c-4de9-9ab3-11cb6fd83225-kube-api-access-q5c7k\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.452366 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-bundle\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.453212 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-bundle\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.454088 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-util\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.484040 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5c7k\" (UniqueName: \"kubernetes.io/projected/a7f6140b-914c-4de9-9ab3-11cb6fd83225-kube-api-access-q5c7k\") pod \"fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:54 crc kubenswrapper[4788]: I1010 15:01:54.579998 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:55 crc kubenswrapper[4788]: I1010 15:01:55.062373 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp"] Oct 10 15:01:55 crc kubenswrapper[4788]: I1010 15:01:55.644875 4788 generic.go:334] "Generic (PLEG): container finished" podID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerID="13e88b6cf22054fda5ee3dadc67e189f914a7b0798940006cc0578244a8f8bc7" exitCode=0 Oct 10 15:01:55 crc kubenswrapper[4788]: I1010 15:01:55.644965 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" event={"ID":"a7f6140b-914c-4de9-9ab3-11cb6fd83225","Type":"ContainerDied","Data":"13e88b6cf22054fda5ee3dadc67e189f914a7b0798940006cc0578244a8f8bc7"} Oct 10 15:01:55 crc kubenswrapper[4788]: I1010 15:01:55.645202 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" event={"ID":"a7f6140b-914c-4de9-9ab3-11cb6fd83225","Type":"ContainerStarted","Data":"81d74f783702a7b094f1f82c8e6fcf166f55af45f71adf885ea473f31b925907"} Oct 10 15:01:56 crc kubenswrapper[4788]: I1010 15:01:56.654837 4788 generic.go:334] "Generic (PLEG): container finished" podID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerID="275b9fb94d1edec33ca83463d7f478e684568a0637f78659dbe9a386eaf5c90c" exitCode=0 Oct 10 15:01:56 crc kubenswrapper[4788]: I1010 15:01:56.654932 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" event={"ID":"a7f6140b-914c-4de9-9ab3-11cb6fd83225","Type":"ContainerDied","Data":"275b9fb94d1edec33ca83463d7f478e684568a0637f78659dbe9a386eaf5c90c"} Oct 10 15:01:57 crc kubenswrapper[4788]: I1010 15:01:57.667336 4788 generic.go:334] "Generic (PLEG): container finished" podID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerID="9bd263ac617b85bae44bb4a4681dc77035e63a00c51442d8b321e646f8d17008" exitCode=0 Oct 10 15:01:57 crc kubenswrapper[4788]: I1010 15:01:57.667384 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" event={"ID":"a7f6140b-914c-4de9-9ab3-11cb6fd83225","Type":"ContainerDied","Data":"9bd263ac617b85bae44bb4a4681dc77035e63a00c51442d8b321e646f8d17008"} Oct 10 15:01:58 crc kubenswrapper[4788]: I1010 15:01:58.947989 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.135367 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-util\") pod \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.135649 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5c7k\" (UniqueName: \"kubernetes.io/projected/a7f6140b-914c-4de9-9ab3-11cb6fd83225-kube-api-access-q5c7k\") pod \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.135708 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-bundle\") pod \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\" (UID: \"a7f6140b-914c-4de9-9ab3-11cb6fd83225\") " Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.136976 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-bundle" (OuterVolumeSpecName: "bundle") pod "a7f6140b-914c-4de9-9ab3-11cb6fd83225" (UID: "a7f6140b-914c-4de9-9ab3-11cb6fd83225"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.146124 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f6140b-914c-4de9-9ab3-11cb6fd83225-kube-api-access-q5c7k" (OuterVolumeSpecName: "kube-api-access-q5c7k") pod "a7f6140b-914c-4de9-9ab3-11cb6fd83225" (UID: "a7f6140b-914c-4de9-9ab3-11cb6fd83225"). InnerVolumeSpecName "kube-api-access-q5c7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.160338 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-util" (OuterVolumeSpecName: "util") pod "a7f6140b-914c-4de9-9ab3-11cb6fd83225" (UID: "a7f6140b-914c-4de9-9ab3-11cb6fd83225"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.238627 4788 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-util\") on node \"crc\" DevicePath \"\"" Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.238714 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5c7k\" (UniqueName: \"kubernetes.io/projected/a7f6140b-914c-4de9-9ab3-11cb6fd83225-kube-api-access-q5c7k\") on node \"crc\" DevicePath \"\"" Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.238749 4788 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a7f6140b-914c-4de9-9ab3-11cb6fd83225-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.705914 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" event={"ID":"a7f6140b-914c-4de9-9ab3-11cb6fd83225","Type":"ContainerDied","Data":"81d74f783702a7b094f1f82c8e6fcf166f55af45f71adf885ea473f31b925907"} Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.706963 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81d74f783702a7b094f1f82c8e6fcf166f55af45f71adf885ea473f31b925907" Oct 10 15:01:59 crc kubenswrapper[4788]: I1010 15:01:59.706018 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.542778 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq"] Oct 10 15:02:01 crc kubenswrapper[4788]: E1010 15:02:01.542980 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerName="extract" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.542993 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerName="extract" Oct 10 15:02:01 crc kubenswrapper[4788]: E1010 15:02:01.543007 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerName="pull" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.543013 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerName="pull" Oct 10 15:02:01 crc kubenswrapper[4788]: E1010 15:02:01.543023 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerName="util" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.543029 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerName="util" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.543158 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f6140b-914c-4de9-9ab3-11cb6fd83225" containerName="extract" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.543733 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.545692 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-dwmx7" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.581619 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq"] Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.673112 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55ltj\" (UniqueName: \"kubernetes.io/projected/70d295ac-34a2-4d46-9d27-27c859268604-kube-api-access-55ltj\") pod \"openstack-operator-controller-operator-65c468ddf8-db6bq\" (UID: \"70d295ac-34a2-4d46-9d27-27c859268604\") " pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.775475 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55ltj\" (UniqueName: \"kubernetes.io/projected/70d295ac-34a2-4d46-9d27-27c859268604-kube-api-access-55ltj\") pod \"openstack-operator-controller-operator-65c468ddf8-db6bq\" (UID: \"70d295ac-34a2-4d46-9d27-27c859268604\") " pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.827021 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55ltj\" (UniqueName: \"kubernetes.io/projected/70d295ac-34a2-4d46-9d27-27c859268604-kube-api-access-55ltj\") pod \"openstack-operator-controller-operator-65c468ddf8-db6bq\" (UID: \"70d295ac-34a2-4d46-9d27-27c859268604\") " pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" Oct 10 15:02:01 crc kubenswrapper[4788]: I1010 15:02:01.875820 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" Oct 10 15:02:02 crc kubenswrapper[4788]: I1010 15:02:02.346317 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq"] Oct 10 15:02:02 crc kubenswrapper[4788]: I1010 15:02:02.737706 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" event={"ID":"70d295ac-34a2-4d46-9d27-27c859268604","Type":"ContainerStarted","Data":"38fcb9f2da71db30a33bc498d4e44c4fb43cbade7ae7921e4ebd9596ed5e3fd3"} Oct 10 15:02:06 crc kubenswrapper[4788]: I1010 15:02:06.782704 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" event={"ID":"70d295ac-34a2-4d46-9d27-27c859268604","Type":"ContainerStarted","Data":"013d57a8faee0b0b24278879da6e7b93848345bf0c9c8cd9d592f9da6756b556"} Oct 10 15:02:08 crc kubenswrapper[4788]: I1010 15:02:08.799453 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" event={"ID":"70d295ac-34a2-4d46-9d27-27c859268604","Type":"ContainerStarted","Data":"86c26ea99bfaea0960f889baefab21ccdb22be739f93f652d2241e6f8cca9c75"} Oct 10 15:02:08 crc kubenswrapper[4788]: I1010 15:02:08.799900 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" Oct 10 15:02:08 crc kubenswrapper[4788]: I1010 15:02:08.851588 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" podStartSLOduration=1.686453158 podStartE2EDuration="7.851568087s" podCreationTimestamp="2025-10-10 15:02:01 +0000 UTC" firstStartedPulling="2025-10-10 15:02:02.359083923 +0000 UTC m=+1024.808799471" lastFinishedPulling="2025-10-10 15:02:08.524198852 +0000 UTC m=+1030.973914400" observedRunningTime="2025-10-10 15:02:08.847480656 +0000 UTC m=+1031.297196204" watchObservedRunningTime="2025-10-10 15:02:08.851568087 +0000 UTC m=+1031.301283635" Oct 10 15:02:11 crc kubenswrapper[4788]: I1010 15:02:11.880406 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-65c468ddf8-db6bq" Oct 10 15:02:29 crc kubenswrapper[4788]: I1010 15:02:29.406972 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:02:29 crc kubenswrapper[4788]: I1010 15:02:29.407875 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.656954 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.658647 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.660953 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qfclg" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.680561 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.689064 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.690810 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.694415 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-66s5s" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.700372 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.739837 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.750180 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.753951 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zjvpc" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.777433 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.783936 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.794940 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-mprhk" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.795729 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqtgp\" (UniqueName: \"kubernetes.io/projected/1d47d678-34c1-41d7-a8cf-6f13199cc741-kube-api-access-kqtgp\") pod \"barbican-operator-controller-manager-64f84fcdbb-k5ls7\" (UID: \"1d47d678-34c1-41d7-a8cf-6f13199cc741\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.796665 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.808046 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.815665 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7qk2t" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.815897 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.840791 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.841891 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.846367 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-kgp58" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.876281 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.898230 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.905715 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh4jk\" (UniqueName: \"kubernetes.io/projected/84cbe7f0-4f13-468c-af71-022960cd0973-kube-api-access-fh4jk\") pod \"designate-operator-controller-manager-687df44cdb-ckw79\" (UID: \"84cbe7f0-4f13-468c-af71-022960cd0973\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.905780 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-l678z"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.905901 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqtgp\" (UniqueName: \"kubernetes.io/projected/1d47d678-34c1-41d7-a8cf-6f13199cc741-kube-api-access-kqtgp\") pod \"barbican-operator-controller-manager-64f84fcdbb-k5ls7\" (UID: \"1d47d678-34c1-41d7-a8cf-6f13199cc741\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.905928 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l9kg\" (UniqueName: \"kubernetes.io/projected/2f9d641b-50d5-4dd9-baec-3c22d23273e1-kube-api-access-6l9kg\") pod \"cinder-operator-controller-manager-59cdc64769-9pbsq\" (UID: \"2f9d641b-50d5-4dd9-baec-3c22d23273e1\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.905968 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msh59\" (UniqueName: \"kubernetes.io/projected/9c8f6b95-2440-4fd6-b0a9-da1b17facf7e-kube-api-access-msh59\") pod \"heat-operator-controller-manager-6d9967f8dd-929g7\" (UID: \"9c8f6b95-2440-4fd6-b0a9-da1b17facf7e\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.907091 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.909171 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nhznq" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.910434 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.927236 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.942080 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-l678z"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.952189 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.953372 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.962929 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-j8sz6" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.965780 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqtgp\" (UniqueName: \"kubernetes.io/projected/1d47d678-34c1-41d7-a8cf-6f13199cc741-kube-api-access-kqtgp\") pod \"barbican-operator-controller-manager-64f84fcdbb-k5ls7\" (UID: \"1d47d678-34c1-41d7-a8cf-6f13199cc741\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.977394 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.978477 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.979204 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.979540 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.983478 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6gfgq" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.983617 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-nqmwq" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.993550 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n"] Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.994552 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.995096 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" Oct 10 15:02:37 crc kubenswrapper[4788]: I1010 15:02:37.995908 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-mpcsc" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.000938 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.004030 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010591 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t98zg\" (UniqueName: \"kubernetes.io/projected/45ccc74a-b207-4512-86fd-07e0495d1378-kube-api-access-t98zg\") pod \"horizon-operator-controller-manager-6d74794d9b-j4xpv\" (UID: \"45ccc74a-b207-4512-86fd-07e0495d1378\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010630 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbzr2\" (UniqueName: \"kubernetes.io/projected/4e656fbd-3ed0-4434-9264-5a08a20607aa-kube-api-access-lbzr2\") pod \"infra-operator-controller-manager-585fc5b659-l678z\" (UID: \"4e656fbd-3ed0-4434-9264-5a08a20607aa\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010651 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdlkx\" (UniqueName: \"kubernetes.io/projected/aae7a82d-e804-4f31-a8be-cc52317be3f1-kube-api-access-xdlkx\") pod \"manila-operator-controller-manager-59578bc799-8l7ct\" (UID: \"aae7a82d-e804-4f31-a8be-cc52317be3f1\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010678 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l9kg\" (UniqueName: \"kubernetes.io/projected/2f9d641b-50d5-4dd9-baec-3c22d23273e1-kube-api-access-6l9kg\") pod \"cinder-operator-controller-manager-59cdc64769-9pbsq\" (UID: \"2f9d641b-50d5-4dd9-baec-3c22d23273e1\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010701 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nltrv\" (UniqueName: \"kubernetes.io/projected/67694f95-9859-40b3-ae66-65cfdc1d17f5-kube-api-access-nltrv\") pod \"keystone-operator-controller-manager-ddb98f99b-4spql\" (UID: \"67694f95-9859-40b3-ae66-65cfdc1d17f5\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010717 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvtt6\" (UniqueName: \"kubernetes.io/projected/682f41df-b648-476f-8a3e-9475501bf290-kube-api-access-jvtt6\") pod \"mariadb-operator-controller-manager-5777b4f897-pkc5n\" (UID: \"682f41df-b648-476f-8a3e-9475501bf290\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010745 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bf8l\" (UniqueName: \"kubernetes.io/projected/8d40607c-b865-4722-93db-94f3fb0b2f42-kube-api-access-8bf8l\") pod \"glance-operator-controller-manager-7bb46cd7d-5v88l\" (UID: \"8d40607c-b865-4722-93db-94f3fb0b2f42\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010768 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msh59\" (UniqueName: \"kubernetes.io/projected/9c8f6b95-2440-4fd6-b0a9-da1b17facf7e-kube-api-access-msh59\") pod \"heat-operator-controller-manager-6d9967f8dd-929g7\" (UID: \"9c8f6b95-2440-4fd6-b0a9-da1b17facf7e\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010785 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e656fbd-3ed0-4434-9264-5a08a20607aa-cert\") pod \"infra-operator-controller-manager-585fc5b659-l678z\" (UID: \"4e656fbd-3ed0-4434-9264-5a08a20607aa\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010822 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh4jk\" (UniqueName: \"kubernetes.io/projected/84cbe7f0-4f13-468c-af71-022960cd0973-kube-api-access-fh4jk\") pod \"designate-operator-controller-manager-687df44cdb-ckw79\" (UID: \"84cbe7f0-4f13-468c-af71-022960cd0973\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.010844 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4dl4\" (UniqueName: \"kubernetes.io/projected/f3166185-7572-4637-b2c6-3b81a1cd79af-kube-api-access-m4dl4\") pod \"ironic-operator-controller-manager-74cb5cbc49-5hstt\" (UID: \"f3166185-7572-4637-b2c6-3b81a1cd79af\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.019162 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.020649 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.023866 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.029094 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.036805 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-vmcqf" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.046574 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l9kg\" (UniqueName: \"kubernetes.io/projected/2f9d641b-50d5-4dd9-baec-3c22d23273e1-kube-api-access-6l9kg\") pod \"cinder-operator-controller-manager-59cdc64769-9pbsq\" (UID: \"2f9d641b-50d5-4dd9-baec-3c22d23273e1\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.046664 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.048027 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.052034 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.057627 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-7d9vd" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.060720 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.063801 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.065211 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.069420 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.069891 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msh59\" (UniqueName: \"kubernetes.io/projected/9c8f6b95-2440-4fd6-b0a9-da1b17facf7e-kube-api-access-msh59\") pod \"heat-operator-controller-manager-6d9967f8dd-929g7\" (UID: \"9c8f6b95-2440-4fd6-b0a9-da1b17facf7e\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.071910 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-c8m4g" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.072127 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.073664 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.076284 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-l9vrm" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.082943 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh4jk\" (UniqueName: \"kubernetes.io/projected/84cbe7f0-4f13-468c-af71-022960cd0973-kube-api-access-fh4jk\") pod \"designate-operator-controller-manager-687df44cdb-ckw79\" (UID: \"84cbe7f0-4f13-468c-af71-022960cd0973\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.083963 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.091682 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.096160 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.096716 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.106384 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.107354 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.108706 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nvpk6" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.134495 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bf8l\" (UniqueName: \"kubernetes.io/projected/8d40607c-b865-4722-93db-94f3fb0b2f42-kube-api-access-8bf8l\") pod \"glance-operator-controller-manager-7bb46cd7d-5v88l\" (UID: \"8d40607c-b865-4722-93db-94f3fb0b2f42\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.134558 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e656fbd-3ed0-4434-9264-5a08a20607aa-cert\") pod \"infra-operator-controller-manager-585fc5b659-l678z\" (UID: \"4e656fbd-3ed0-4434-9264-5a08a20607aa\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.134594 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqqqj\" (UniqueName: \"kubernetes.io/projected/d24a6fbc-3938-4612-a928-c0d97cba7fa8-kube-api-access-xqqqj\") pod \"ovn-operator-controller-manager-869cc7797f-kbw8c\" (UID: \"d24a6fbc-3938-4612-a928-c0d97cba7fa8\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.134621 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vv2v\" (UniqueName: \"kubernetes.io/projected/6791183a-bdf7-4ce2-b888-cd7e1b70ab80-kube-api-access-4vv2v\") pod \"octavia-operator-controller-manager-6d7c7ddf95-x2mn7\" (UID: \"6791183a-bdf7-4ce2-b888-cd7e1b70ab80\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" Oct 10 15:02:38 crc kubenswrapper[4788]: E1010 15:02:38.136004 4788 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 10 15:02:38 crc kubenswrapper[4788]: E1010 15:02:38.136109 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e656fbd-3ed0-4434-9264-5a08a20607aa-cert podName:4e656fbd-3ed0-4434-9264-5a08a20607aa nodeName:}" failed. No retries permitted until 2025-10-10 15:02:38.636081501 +0000 UTC m=+1061.085797039 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4e656fbd-3ed0-4434-9264-5a08a20607aa-cert") pod "infra-operator-controller-manager-585fc5b659-l678z" (UID: "4e656fbd-3ed0-4434-9264-5a08a20607aa") : secret "infra-operator-webhook-server-cert" not found Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.136736 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.136949 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtrrw\" (UniqueName: \"kubernetes.io/projected/07597e46-5cc8-495d-95a1-f94235f2f9ed-kube-api-access-wtrrw\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz\" (UID: \"07597e46-5cc8-495d-95a1-f94235f2f9ed\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.137186 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz\" (UID: \"07597e46-5cc8-495d-95a1-f94235f2f9ed\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.137216 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4dl4\" (UniqueName: \"kubernetes.io/projected/f3166185-7572-4637-b2c6-3b81a1cd79af-kube-api-access-m4dl4\") pod \"ironic-operator-controller-manager-74cb5cbc49-5hstt\" (UID: \"f3166185-7572-4637-b2c6-3b81a1cd79af\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.137280 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ltqx\" (UniqueName: \"kubernetes.io/projected/6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0-kube-api-access-5ltqx\") pod \"nova-operator-controller-manager-57bb74c7bf-2cwz7\" (UID: \"6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.137352 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t98zg\" (UniqueName: \"kubernetes.io/projected/45ccc74a-b207-4512-86fd-07e0495d1378-kube-api-access-t98zg\") pod \"horizon-operator-controller-manager-6d74794d9b-j4xpv\" (UID: \"45ccc74a-b207-4512-86fd-07e0495d1378\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.137413 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbzr2\" (UniqueName: \"kubernetes.io/projected/4e656fbd-3ed0-4434-9264-5a08a20607aa-kube-api-access-lbzr2\") pod \"infra-operator-controller-manager-585fc5b659-l678z\" (UID: \"4e656fbd-3ed0-4434-9264-5a08a20607aa\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.137440 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdlkx\" (UniqueName: \"kubernetes.io/projected/aae7a82d-e804-4f31-a8be-cc52317be3f1-kube-api-access-xdlkx\") pod \"manila-operator-controller-manager-59578bc799-8l7ct\" (UID: \"aae7a82d-e804-4f31-a8be-cc52317be3f1\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.137484 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nltrv\" (UniqueName: \"kubernetes.io/projected/67694f95-9859-40b3-ae66-65cfdc1d17f5-kube-api-access-nltrv\") pod \"keystone-operator-controller-manager-ddb98f99b-4spql\" (UID: \"67694f95-9859-40b3-ae66-65cfdc1d17f5\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.137510 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvtt6\" (UniqueName: \"kubernetes.io/projected/682f41df-b648-476f-8a3e-9475501bf290-kube-api-access-jvtt6\") pod \"mariadb-operator-controller-manager-5777b4f897-pkc5n\" (UID: \"682f41df-b648-476f-8a3e-9475501bf290\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.156370 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.156552 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.164902 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-9lts6" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.170629 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.174754 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbzr2\" (UniqueName: \"kubernetes.io/projected/4e656fbd-3ed0-4434-9264-5a08a20607aa-kube-api-access-lbzr2\") pod \"infra-operator-controller-manager-585fc5b659-l678z\" (UID: \"4e656fbd-3ed0-4434-9264-5a08a20607aa\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.175207 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nltrv\" (UniqueName: \"kubernetes.io/projected/67694f95-9859-40b3-ae66-65cfdc1d17f5-kube-api-access-nltrv\") pod \"keystone-operator-controller-manager-ddb98f99b-4spql\" (UID: \"67694f95-9859-40b3-ae66-65cfdc1d17f5\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.181964 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdlkx\" (UniqueName: \"kubernetes.io/projected/aae7a82d-e804-4f31-a8be-cc52317be3f1-kube-api-access-xdlkx\") pod \"manila-operator-controller-manager-59578bc799-8l7ct\" (UID: \"aae7a82d-e804-4f31-a8be-cc52317be3f1\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.178708 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.182095 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bf8l\" (UniqueName: \"kubernetes.io/projected/8d40607c-b865-4722-93db-94f3fb0b2f42-kube-api-access-8bf8l\") pod \"glance-operator-controller-manager-7bb46cd7d-5v88l\" (UID: \"8d40607c-b865-4722-93db-94f3fb0b2f42\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.182593 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.184328 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-fx2wb" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.186080 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvtt6\" (UniqueName: \"kubernetes.io/projected/682f41df-b648-476f-8a3e-9475501bf290-kube-api-access-jvtt6\") pod \"mariadb-operator-controller-manager-5777b4f897-pkc5n\" (UID: \"682f41df-b648-476f-8a3e-9475501bf290\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.190167 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4dl4\" (UniqueName: \"kubernetes.io/projected/f3166185-7572-4637-b2c6-3b81a1cd79af-kube-api-access-m4dl4\") pod \"ironic-operator-controller-manager-74cb5cbc49-5hstt\" (UID: \"f3166185-7572-4637-b2c6-3b81a1cd79af\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.190313 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t98zg\" (UniqueName: \"kubernetes.io/projected/45ccc74a-b207-4512-86fd-07e0495d1378-kube-api-access-t98zg\") pod \"horizon-operator-controller-manager-6d74794d9b-j4xpv\" (UID: \"45ccc74a-b207-4512-86fd-07e0495d1378\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.190654 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.214894 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.216179 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.218607 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4dr4n" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244451 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz\" (UID: \"07597e46-5cc8-495d-95a1-f94235f2f9ed\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244489 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp5sj\" (UniqueName: \"kubernetes.io/projected/22f63d73-f623-4b6a-83d7-dff55231ecc7-kube-api-access-bp5sj\") pod \"telemetry-operator-controller-manager-578874c84d-jvfjk\" (UID: \"22f63d73-f623-4b6a-83d7-dff55231ecc7\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244516 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ltqx\" (UniqueName: \"kubernetes.io/projected/6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0-kube-api-access-5ltqx\") pod \"nova-operator-controller-manager-57bb74c7bf-2cwz7\" (UID: \"6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244569 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7srb\" (UniqueName: \"kubernetes.io/projected/38f6b32d-a5f6-402f-8070-eb1252c7d669-kube-api-access-j7srb\") pod \"swift-operator-controller-manager-5f4d5dfdc6-2z2xl\" (UID: \"38f6b32d-a5f6-402f-8070-eb1252c7d669\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244613 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqqqj\" (UniqueName: \"kubernetes.io/projected/d24a6fbc-3938-4612-a928-c0d97cba7fa8-kube-api-access-xqqqj\") pod \"ovn-operator-controller-manager-869cc7797f-kbw8c\" (UID: \"d24a6fbc-3938-4612-a928-c0d97cba7fa8\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244742 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vv2v\" (UniqueName: \"kubernetes.io/projected/6791183a-bdf7-4ce2-b888-cd7e1b70ab80-kube-api-access-4vv2v\") pod \"octavia-operator-controller-manager-6d7c7ddf95-x2mn7\" (UID: \"6791183a-bdf7-4ce2-b888-cd7e1b70ab80\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244766 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cs74\" (UniqueName: \"kubernetes.io/projected/41d202a5-eddf-4ee7-9b9a-9edf7905d51f-kube-api-access-8cs74\") pod \"neutron-operator-controller-manager-797d478b46-mvw9d\" (UID: \"41d202a5-eddf-4ee7-9b9a-9edf7905d51f\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244790 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtrrw\" (UniqueName: \"kubernetes.io/projected/07597e46-5cc8-495d-95a1-f94235f2f9ed-kube-api-access-wtrrw\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz\" (UID: \"07597e46-5cc8-495d-95a1-f94235f2f9ed\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.244820 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcx22\" (UniqueName: \"kubernetes.io/projected/d99a1777-7ea0-4072-b086-aabc93f57f71-kube-api-access-pcx22\") pod \"placement-operator-controller-manager-664664cb68-4gq6b\" (UID: \"d99a1777-7ea0-4072-b086-aabc93f57f71\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" Oct 10 15:02:38 crc kubenswrapper[4788]: E1010 15:02:38.244940 4788 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 15:02:38 crc kubenswrapper[4788]: E1010 15:02:38.244981 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert podName:07597e46-5cc8-495d-95a1-f94235f2f9ed nodeName:}" failed. No retries permitted until 2025-10-10 15:02:38.744966495 +0000 UTC m=+1061.194682043 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" (UID: "07597e46-5cc8-495d-95a1-f94235f2f9ed") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.279845 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.280307 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.281960 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.290193 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-b6kxl" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.293104 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ltqx\" (UniqueName: \"kubernetes.io/projected/6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0-kube-api-access-5ltqx\") pod \"nova-operator-controller-manager-57bb74c7bf-2cwz7\" (UID: \"6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.294990 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.297240 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqqqj\" (UniqueName: \"kubernetes.io/projected/d24a6fbc-3938-4612-a928-c0d97cba7fa8-kube-api-access-xqqqj\") pod \"ovn-operator-controller-manager-869cc7797f-kbw8c\" (UID: \"d24a6fbc-3938-4612-a928-c0d97cba7fa8\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.313329 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.315226 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.322665 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtrrw\" (UniqueName: \"kubernetes.io/projected/07597e46-5cc8-495d-95a1-f94235f2f9ed-kube-api-access-wtrrw\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz\" (UID: \"07597e46-5cc8-495d-95a1-f94235f2f9ed\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.341249 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-hcq88"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.342698 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.348439 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.349921 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7srb\" (UniqueName: \"kubernetes.io/projected/38f6b32d-a5f6-402f-8070-eb1252c7d669-kube-api-access-j7srb\") pod \"swift-operator-controller-manager-5f4d5dfdc6-2z2xl\" (UID: \"38f6b32d-a5f6-402f-8070-eb1252c7d669\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.350014 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cs74\" (UniqueName: \"kubernetes.io/projected/41d202a5-eddf-4ee7-9b9a-9edf7905d51f-kube-api-access-8cs74\") pod \"neutron-operator-controller-manager-797d478b46-mvw9d\" (UID: \"41d202a5-eddf-4ee7-9b9a-9edf7905d51f\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.350175 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcx22\" (UniqueName: \"kubernetes.io/projected/d99a1777-7ea0-4072-b086-aabc93f57f71-kube-api-access-pcx22\") pod \"placement-operator-controller-manager-664664cb68-4gq6b\" (UID: \"d99a1777-7ea0-4072-b086-aabc93f57f71\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.350229 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp5sj\" (UniqueName: \"kubernetes.io/projected/22f63d73-f623-4b6a-83d7-dff55231ecc7-kube-api-access-bp5sj\") pod \"telemetry-operator-controller-manager-578874c84d-jvfjk\" (UID: \"22f63d73-f623-4b6a-83d7-dff55231ecc7\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.358847 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-zbpgb" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.358981 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vv2v\" (UniqueName: \"kubernetes.io/projected/6791183a-bdf7-4ce2-b888-cd7e1b70ab80-kube-api-access-4vv2v\") pod \"octavia-operator-controller-manager-6d7c7ddf95-x2mn7\" (UID: \"6791183a-bdf7-4ce2-b888-cd7e1b70ab80\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.362176 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-hcq88"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.375506 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cs74\" (UniqueName: \"kubernetes.io/projected/41d202a5-eddf-4ee7-9b9a-9edf7905d51f-kube-api-access-8cs74\") pod \"neutron-operator-controller-manager-797d478b46-mvw9d\" (UID: \"41d202a5-eddf-4ee7-9b9a-9edf7905d51f\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.376752 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7srb\" (UniqueName: \"kubernetes.io/projected/38f6b32d-a5f6-402f-8070-eb1252c7d669-kube-api-access-j7srb\") pod \"swift-operator-controller-manager-5f4d5dfdc6-2z2xl\" (UID: \"38f6b32d-a5f6-402f-8070-eb1252c7d669\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.388865 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp5sj\" (UniqueName: \"kubernetes.io/projected/22f63d73-f623-4b6a-83d7-dff55231ecc7-kube-api-access-bp5sj\") pod \"telemetry-operator-controller-manager-578874c84d-jvfjk\" (UID: \"22f63d73-f623-4b6a-83d7-dff55231ecc7\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.395650 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcx22\" (UniqueName: \"kubernetes.io/projected/d99a1777-7ea0-4072-b086-aabc93f57f71-kube-api-access-pcx22\") pod \"placement-operator-controller-manager-664664cb68-4gq6b\" (UID: \"d99a1777-7ea0-4072-b086-aabc93f57f71\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.404178 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.410430 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.433295 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.447105 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.447928 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.460027 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.461305 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf6nm\" (UniqueName: \"kubernetes.io/projected/ebfc2c03-b35c-41a7-9679-11677bca30b2-kube-api-access-kf6nm\") pod \"test-operator-controller-manager-ffcdd6c94-7tlj8\" (UID: \"ebfc2c03-b35c-41a7-9679-11677bca30b2\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.461397 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzsfq\" (UniqueName: \"kubernetes.io/projected/c2dbece7-c07a-4223-98a2-bb96f93b9e3c-kube-api-access-gzsfq\") pod \"watcher-operator-controller-manager-646675d848-hcq88\" (UID: \"c2dbece7-c07a-4223-98a2-bb96f93b9e3c\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.462233 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.469205 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.469345 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-h6ggg" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.469649 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.493624 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.496214 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.511390 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.543516 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.549053 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.551576 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.555077 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-wr5cc" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.555919 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.559368 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.566793 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf6nm\" (UniqueName: \"kubernetes.io/projected/ebfc2c03-b35c-41a7-9679-11677bca30b2-kube-api-access-kf6nm\") pod \"test-operator-controller-manager-ffcdd6c94-7tlj8\" (UID: \"ebfc2c03-b35c-41a7-9679-11677bca30b2\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.566869 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzsfq\" (UniqueName: \"kubernetes.io/projected/c2dbece7-c07a-4223-98a2-bb96f93b9e3c-kube-api-access-gzsfq\") pod \"watcher-operator-controller-manager-646675d848-hcq88\" (UID: \"c2dbece7-c07a-4223-98a2-bb96f93b9e3c\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.587004 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf6nm\" (UniqueName: \"kubernetes.io/projected/ebfc2c03-b35c-41a7-9679-11677bca30b2-kube-api-access-kf6nm\") pod \"test-operator-controller-manager-ffcdd6c94-7tlj8\" (UID: \"ebfc2c03-b35c-41a7-9679-11677bca30b2\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.589662 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzsfq\" (UniqueName: \"kubernetes.io/projected/c2dbece7-c07a-4223-98a2-bb96f93b9e3c-kube-api-access-gzsfq\") pod \"watcher-operator-controller-manager-646675d848-hcq88\" (UID: \"c2dbece7-c07a-4223-98a2-bb96f93b9e3c\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.591434 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.595257 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.671534 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e656fbd-3ed0-4434-9264-5a08a20607aa-cert\") pod \"infra-operator-controller-manager-585fc5b659-l678z\" (UID: \"4e656fbd-3ed0-4434-9264-5a08a20607aa\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.671628 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7dxl\" (UniqueName: \"kubernetes.io/projected/10351a76-d81b-4d40-935e-2dc3eb1b3960-kube-api-access-j7dxl\") pod \"openstack-operator-controller-manager-749665c6b7-qx9vb\" (UID: \"10351a76-d81b-4d40-935e-2dc3eb1b3960\") " pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.671659 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxmpd\" (UniqueName: \"kubernetes.io/projected/7940c932-25d0-4f02-bcd5-e3441aa026c4-kube-api-access-xxmpd\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv\" (UID: \"7940c932-25d0-4f02-bcd5-e3441aa026c4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.671688 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10351a76-d81b-4d40-935e-2dc3eb1b3960-cert\") pod \"openstack-operator-controller-manager-749665c6b7-qx9vb\" (UID: \"10351a76-d81b-4d40-935e-2dc3eb1b3960\") " pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.684239 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4e656fbd-3ed0-4434-9264-5a08a20607aa-cert\") pod \"infra-operator-controller-manager-585fc5b659-l678z\" (UID: \"4e656fbd-3ed0-4434-9264-5a08a20607aa\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.775621 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz\" (UID: \"07597e46-5cc8-495d-95a1-f94235f2f9ed\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.776010 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7dxl\" (UniqueName: \"kubernetes.io/projected/10351a76-d81b-4d40-935e-2dc3eb1b3960-kube-api-access-j7dxl\") pod \"openstack-operator-controller-manager-749665c6b7-qx9vb\" (UID: \"10351a76-d81b-4d40-935e-2dc3eb1b3960\") " pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.776037 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxmpd\" (UniqueName: \"kubernetes.io/projected/7940c932-25d0-4f02-bcd5-e3441aa026c4-kube-api-access-xxmpd\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv\" (UID: \"7940c932-25d0-4f02-bcd5-e3441aa026c4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.776069 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10351a76-d81b-4d40-935e-2dc3eb1b3960-cert\") pod \"openstack-operator-controller-manager-749665c6b7-qx9vb\" (UID: \"10351a76-d81b-4d40-935e-2dc3eb1b3960\") " pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:38 crc kubenswrapper[4788]: E1010 15:02:38.776220 4788 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 15:02:38 crc kubenswrapper[4788]: E1010 15:02:38.776290 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert podName:07597e46-5cc8-495d-95a1-f94235f2f9ed nodeName:}" failed. No retries permitted until 2025-10-10 15:02:39.776271356 +0000 UTC m=+1062.225986904 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" (UID: "07597e46-5cc8-495d-95a1-f94235f2f9ed") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 10 15:02:38 crc kubenswrapper[4788]: E1010 15:02:38.776227 4788 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 10 15:02:38 crc kubenswrapper[4788]: E1010 15:02:38.778135 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10351a76-d81b-4d40-935e-2dc3eb1b3960-cert podName:10351a76-d81b-4d40-935e-2dc3eb1b3960 nodeName:}" failed. No retries permitted until 2025-10-10 15:02:39.278118005 +0000 UTC m=+1061.727833553 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/10351a76-d81b-4d40-935e-2dc3eb1b3960-cert") pod "openstack-operator-controller-manager-749665c6b7-qx9vb" (UID: "10351a76-d81b-4d40-935e-2dc3eb1b3960") : secret "webhook-server-cert" not found Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.802446 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxmpd\" (UniqueName: \"kubernetes.io/projected/7940c932-25d0-4f02-bcd5-e3441aa026c4-kube-api-access-xxmpd\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv\" (UID: \"7940c932-25d0-4f02-bcd5-e3441aa026c4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.804659 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7dxl\" (UniqueName: \"kubernetes.io/projected/10351a76-d81b-4d40-935e-2dc3eb1b3960-kube-api-access-j7dxl\") pod \"openstack-operator-controller-manager-749665c6b7-qx9vb\" (UID: \"10351a76-d81b-4d40-935e-2dc3eb1b3960\") " pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.851859 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.867571 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.868000 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.943903 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv" Oct 10 15:02:38 crc kubenswrapper[4788]: W1010 15:02:38.953767 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d47d678_34c1_41d7_a8cf_6f13199cc741.slice/crio-63c9932308d7deaca7878ebcc59261aabea9d871a6fb6da3f79a8677a594b555 WatchSource:0}: Error finding container 63c9932308d7deaca7878ebcc59261aabea9d871a6fb6da3f79a8677a594b555: Status 404 returned error can't find the container with id 63c9932308d7deaca7878ebcc59261aabea9d871a6fb6da3f79a8677a594b555 Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.956165 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7"] Oct 10 15:02:38 crc kubenswrapper[4788]: I1010 15:02:38.984314 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.071594 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" event={"ID":"9c8f6b95-2440-4fd6-b0a9-da1b17facf7e","Type":"ContainerStarted","Data":"fe0332e2e745babf22266053fe1c616c35d64c427744113a2e2788a33de7184e"} Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.073877 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" event={"ID":"1d47d678-34c1-41d7-a8cf-6f13199cc741","Type":"ContainerStarted","Data":"63c9932308d7deaca7878ebcc59261aabea9d871a6fb6da3f79a8677a594b555"} Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.159474 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.183073 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.268037 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.283215 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.284789 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10351a76-d81b-4d40-935e-2dc3eb1b3960-cert\") pod \"openstack-operator-controller-manager-749665c6b7-qx9vb\" (UID: \"10351a76-d81b-4d40-935e-2dc3eb1b3960\") " pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.295880 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10351a76-d81b-4d40-935e-2dc3eb1b3960-cert\") pod \"openstack-operator-controller-manager-749665c6b7-qx9vb\" (UID: \"10351a76-d81b-4d40-935e-2dc3eb1b3960\") " pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.442584 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.455342 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.457774 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk"] Oct 10 15:02:39 crc kubenswrapper[4788]: W1010 15:02:39.466696 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d40607c_b865_4722_93db_94f3fb0b2f42.slice/crio-707a2301082fe0da5daae4ec81390ff29cb85e2878aece3c2146cda0061c748d WatchSource:0}: Error finding container 707a2301082fe0da5daae4ec81390ff29cb85e2878aece3c2146cda0061c748d: Status 404 returned error can't find the container with id 707a2301082fe0da5daae4ec81390ff29cb85e2878aece3c2146cda0061c748d Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.521338 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.750859 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.789625 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7"] Oct 10 15:02:39 crc kubenswrapper[4788]: W1010 15:02:39.789855 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6791183a_bdf7_4ce2_b888_cd7e1b70ab80.slice/crio-32d1081b0a91c86605b70d0e0d438d2a7b2f2d0adef15e64a01494eb17e8f883 WatchSource:0}: Error finding container 32d1081b0a91c86605b70d0e0d438d2a7b2f2d0adef15e64a01494eb17e8f883: Status 404 returned error can't find the container with id 32d1081b0a91c86605b70d0e0d438d2a7b2f2d0adef15e64a01494eb17e8f883 Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.800723 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz\" (UID: \"07597e46-5cc8-495d-95a1-f94235f2f9ed\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.801003 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.804921 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07597e46-5cc8-495d-95a1-f94235f2f9ed-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz\" (UID: \"07597e46-5cc8-495d-95a1-f94235f2f9ed\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.809732 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv"] Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.936265 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:39 crc kubenswrapper[4788]: I1010 15:02:39.962690 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv"] Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.013630 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7"] Oct 10 15:02:40 crc kubenswrapper[4788]: W1010 15:02:40.019414 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7940c932_25d0_4f02_bcd5_e3441aa026c4.slice/crio-1b6fb3267cbff5d9938440229cc61089f6fd70b174a030672cf51030c5110775 WatchSource:0}: Error finding container 1b6fb3267cbff5d9938440229cc61089f6fd70b174a030672cf51030c5110775: Status 404 returned error can't find the container with id 1b6fb3267cbff5d9938440229cc61089f6fd70b174a030672cf51030c5110775 Oct 10 15:02:40 crc kubenswrapper[4788]: W1010 15:02:40.029443 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ddf56dc_44b6_49a7_9bf6_91362cdd3ea0.slice/crio-cfac930b20af03e2559637559c645aeeca0b2a427c3d361b3086280a5f71d59a WatchSource:0}: Error finding container cfac930b20af03e2559637559c645aeeca0b2a427c3d361b3086280a5f71d59a: Status 404 returned error can't find the container with id cfac930b20af03e2559637559c645aeeca0b2a427c3d361b3086280a5f71d59a Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.030934 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-hcq88"] Oct 10 15:02:40 crc kubenswrapper[4788]: W1010 15:02:40.036586 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod682f41df_b648_476f_8a3e_9475501bf290.slice/crio-9484d9e94c0dc0255bbb2146ccfe9854a02672b1c7c8b3d5b0c9c11d7ce8827b WatchSource:0}: Error finding container 9484d9e94c0dc0255bbb2146ccfe9854a02672b1c7c8b3d5b0c9c11d7ce8827b: Status 404 returned error can't find the container with id 9484d9e94c0dc0255bbb2146ccfe9854a02672b1c7c8b3d5b0c9c11d7ce8827b Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.037549 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5ltqx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-57bb74c7bf-2cwz7_openstack-operators(6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.039603 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jvtt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-5777b4f897-pkc5n_openstack-operators(682f41df-b648-476f-8a3e-9475501bf290): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.042509 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b"] Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.047498 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pcx22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-664664cb68-4gq6b_openstack-operators(d99a1777-7ea0-4072-b086-aabc93f57f71): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.049158 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n"] Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.056488 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-l678z"] Oct 10 15:02:40 crc kubenswrapper[4788]: W1010 15:02:40.072096 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2dbece7_c07a_4223_98a2_bb96f93b9e3c.slice/crio-3d3abb5594911e1df32be3c5a25d82d30c95c27064a3a357584490778607fbae WatchSource:0}: Error finding container 3d3abb5594911e1df32be3c5a25d82d30c95c27064a3a357584490778607fbae: Status 404 returned error can't find the container with id 3d3abb5594911e1df32be3c5a25d82d30c95c27064a3a357584490778607fbae Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.076381 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gzsfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-646675d848-hcq88_openstack-operators(c2dbece7-c07a-4223-98a2-bb96f93b9e3c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 15:02:40 crc kubenswrapper[4788]: W1010 15:02:40.080387 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e656fbd_3ed0_4434_9264_5a08a20607aa.slice/crio-160ee70f24341e8bd0fec03600eaf5872cdfa566af67957a328e1b478448c00c WatchSource:0}: Error finding container 160ee70f24341e8bd0fec03600eaf5872cdfa566af67957a328e1b478448c00c: Status 404 returned error can't find the container with id 160ee70f24341e8bd0fec03600eaf5872cdfa566af67957a328e1b478448c00c Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.082831 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" event={"ID":"22f63d73-f623-4b6a-83d7-dff55231ecc7","Type":"ContainerStarted","Data":"7bc59c119bf37a6507af6279b196274e7670dd7054b036e41365ff06f6e2c7a6"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.085106 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" event={"ID":"c2dbece7-c07a-4223-98a2-bb96f93b9e3c","Type":"ContainerStarted","Data":"3d3abb5594911e1df32be3c5a25d82d30c95c27064a3a357584490778607fbae"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.087391 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" event={"ID":"682f41df-b648-476f-8a3e-9475501bf290","Type":"ContainerStarted","Data":"9484d9e94c0dc0255bbb2146ccfe9854a02672b1c7c8b3d5b0c9c11d7ce8827b"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.088765 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" event={"ID":"6791183a-bdf7-4ce2-b888-cd7e1b70ab80","Type":"ContainerStarted","Data":"32d1081b0a91c86605b70d0e0d438d2a7b2f2d0adef15e64a01494eb17e8f883"} Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.093881 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lbzr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-585fc5b659-l678z_openstack-operators(4e656fbd-3ed0-4434-9264-5a08a20607aa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.094216 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" event={"ID":"6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0","Type":"ContainerStarted","Data":"cfac930b20af03e2559637559c645aeeca0b2a427c3d361b3086280a5f71d59a"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.097522 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" event={"ID":"45ccc74a-b207-4512-86fd-07e0495d1378","Type":"ContainerStarted","Data":"5981db884da4cb65c13837ba90b3fdbd8fca0842ee9c11278cf30b96aa73b433"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.100333 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" event={"ID":"84cbe7f0-4f13-468c-af71-022960cd0973","Type":"ContainerStarted","Data":"febeff3c84932e0f02e2d19e8156c57000ea37bbff0c71e12e2b406d16df7977"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.125257 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8"] Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.133410 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv" event={"ID":"7940c932-25d0-4f02-bcd5-e3441aa026c4","Type":"ContainerStarted","Data":"1b6fb3267cbff5d9938440229cc61089f6fd70b174a030672cf51030c5110775"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.144542 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" event={"ID":"d24a6fbc-3938-4612-a928-c0d97cba7fa8","Type":"ContainerStarted","Data":"6952453509d0374186ab29f12d2d2da799fcbe01f7e7008a9c4ec2fc58b6084f"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.149246 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" event={"ID":"38f6b32d-a5f6-402f-8070-eb1252c7d669","Type":"ContainerStarted","Data":"614104be582049247e457b7fd14ceef13d10fdaa260dcb72deb2a86af72790f4"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.152379 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" event={"ID":"f3166185-7572-4637-b2c6-3b81a1cd79af","Type":"ContainerStarted","Data":"99d1da0c77022ec888480913a0ce290ad0c7ad39d8edf10f2914c6c120ffffb3"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.155912 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" event={"ID":"d99a1777-7ea0-4072-b086-aabc93f57f71","Type":"ContainerStarted","Data":"e7213dbfc3a5efde2c854eda941276b82622be02c89c27d622226f68cdc2edee"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.159725 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" event={"ID":"2f9d641b-50d5-4dd9-baec-3c22d23273e1","Type":"ContainerStarted","Data":"e9f83930a2b4f8a0c954557082fa0c2f82e96ed438e8c55c1d600132e0de427b"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.164890 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" event={"ID":"41d202a5-eddf-4ee7-9b9a-9edf7905d51f","Type":"ContainerStarted","Data":"666492dec7656509660c31e055e24683429c8938b3e935ca9601c0d62a7e966b"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.194967 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" event={"ID":"8d40607c-b865-4722-93db-94f3fb0b2f42","Type":"ContainerStarted","Data":"707a2301082fe0da5daae4ec81390ff29cb85e2878aece3c2146cda0061c748d"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.197853 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" event={"ID":"aae7a82d-e804-4f31-a8be-cc52317be3f1","Type":"ContainerStarted","Data":"fbab4fb2e7f7b5ebe89009f0778a1ade7b361e7cf0936f9223db814827990121"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.204260 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" event={"ID":"67694f95-9859-40b3-ae66-65cfdc1d17f5","Type":"ContainerStarted","Data":"7c952b33ecf27dee359f7f370019efbb1803cb4323e9ea250de902e6ebedbbd3"} Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.231157 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb"] Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.339539 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" podUID="d99a1777-7ea0-4072-b086-aabc93f57f71" Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.356048 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" podUID="6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0" Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.391568 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" podUID="682f41df-b648-476f-8a3e-9475501bf290" Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.474897 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" podUID="4e656fbd-3ed0-4434-9264-5a08a20607aa" Oct 10 15:02:40 crc kubenswrapper[4788]: E1010 15:02:40.512379 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" podUID="c2dbece7-c07a-4223-98a2-bb96f93b9e3c" Oct 10 15:02:40 crc kubenswrapper[4788]: I1010 15:02:40.569062 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz"] Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.271179 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" event={"ID":"07597e46-5cc8-495d-95a1-f94235f2f9ed","Type":"ContainerStarted","Data":"165b10621bd55ad92e9f373b746d0a38baaa533b40882923a355070a752edc4c"} Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.279755 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" event={"ID":"6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0","Type":"ContainerStarted","Data":"d03107d9dd6fb33385b2e798d99d0c3c5ad4aa78d01dba05e24423a72b207d1d"} Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.283743 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" event={"ID":"c2dbece7-c07a-4223-98a2-bb96f93b9e3c","Type":"ContainerStarted","Data":"e68deb259501d13b2b62e4872774f708d5f93a37ab72f0e2bdfdc2ce234c5e59"} Oct 10 15:02:41 crc kubenswrapper[4788]: E1010 15:02:41.288979 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd\\\"\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" podUID="6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0" Oct 10 15:02:41 crc kubenswrapper[4788]: E1010 15:02:41.288997 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" podUID="c2dbece7-c07a-4223-98a2-bb96f93b9e3c" Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.290185 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" event={"ID":"ebfc2c03-b35c-41a7-9679-11677bca30b2","Type":"ContainerStarted","Data":"9cf62a7a8a310b251920ab0be424b7ab5a4647ae2d78aa8417f153618554659f"} Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.307294 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" event={"ID":"682f41df-b648-476f-8a3e-9475501bf290","Type":"ContainerStarted","Data":"d3f752afd70b10d160de04ba7a95da0a0351d9425fc8ba2e9f5528faa615ba7f"} Oct 10 15:02:41 crc kubenswrapper[4788]: E1010 15:02:41.312876 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" podUID="682f41df-b648-476f-8a3e-9475501bf290" Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.314969 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" event={"ID":"d99a1777-7ea0-4072-b086-aabc93f57f71","Type":"ContainerStarted","Data":"1a88c159ec8a1853b7a99a72f78c254796a3298a343b4519e38044e06e0b08ea"} Oct 10 15:02:41 crc kubenswrapper[4788]: E1010 15:02:41.319716 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" podUID="d99a1777-7ea0-4072-b086-aabc93f57f71" Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.328692 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" event={"ID":"10351a76-d81b-4d40-935e-2dc3eb1b3960","Type":"ContainerStarted","Data":"3af7c227205b7dcc7389950323dde8b973aea7205d8828cd59e9b8c03d6d1823"} Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.328757 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" event={"ID":"10351a76-d81b-4d40-935e-2dc3eb1b3960","Type":"ContainerStarted","Data":"e9289352ced9d8aa682cd80e4866d4f779bc4f71b9a478f315d738d6ce37acb0"} Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.328772 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" event={"ID":"10351a76-d81b-4d40-935e-2dc3eb1b3960","Type":"ContainerStarted","Data":"1c3b4d0b6bf7eff158f32366e2836b42696f72ffa748e86b51faa9de9a94365e"} Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.332785 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.363732 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" event={"ID":"4e656fbd-3ed0-4434-9264-5a08a20607aa","Type":"ContainerStarted","Data":"9bf6ed6dccab3683779287f6d7146a9dd408b7b81769072bacdb0a58cc31e754"} Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.363863 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" event={"ID":"4e656fbd-3ed0-4434-9264-5a08a20607aa","Type":"ContainerStarted","Data":"160ee70f24341e8bd0fec03600eaf5872cdfa566af67957a328e1b478448c00c"} Oct 10 15:02:41 crc kubenswrapper[4788]: E1010 15:02:41.366107 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" podUID="4e656fbd-3ed0-4434-9264-5a08a20607aa" Oct 10 15:02:41 crc kubenswrapper[4788]: I1010 15:02:41.429763 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" podStartSLOduration=3.429743295 podStartE2EDuration="3.429743295s" podCreationTimestamp="2025-10-10 15:02:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:02:41.409396334 +0000 UTC m=+1063.859111882" watchObservedRunningTime="2025-10-10 15:02:41.429743295 +0000 UTC m=+1063.879458843" Oct 10 15:02:42 crc kubenswrapper[4788]: E1010 15:02:42.383868 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" podUID="c2dbece7-c07a-4223-98a2-bb96f93b9e3c" Oct 10 15:02:42 crc kubenswrapper[4788]: E1010 15:02:42.385858 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" podUID="682f41df-b648-476f-8a3e-9475501bf290" Oct 10 15:02:42 crc kubenswrapper[4788]: E1010 15:02:42.385964 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" podUID="4e656fbd-3ed0-4434-9264-5a08a20607aa" Oct 10 15:02:42 crc kubenswrapper[4788]: E1010 15:02:42.386021 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" podUID="d99a1777-7ea0-4072-b086-aabc93f57f71" Oct 10 15:02:42 crc kubenswrapper[4788]: E1010 15:02:42.386864 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd\\\"\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" podUID="6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0" Oct 10 15:02:49 crc kubenswrapper[4788]: I1010 15:02:49.572970 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-749665c6b7-qx9vb" Oct 10 15:02:52 crc kubenswrapper[4788]: E1010 15:02:52.172805 4788 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14" Oct 10 15:02:52 crc kubenswrapper[4788]: E1010 15:02:52.173426 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xqqqj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-869cc7797f-kbw8c_openstack-operators(d24a6fbc-3938-4612-a928-c0d97cba7fa8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 15:02:52 crc kubenswrapper[4788]: E1010 15:02:52.679983 4788 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960" Oct 10 15:02:52 crc kubenswrapper[4788]: E1010 15:02:52.680408 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m4dl4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-74cb5cbc49-5hstt_openstack-operators(f3166185-7572-4637-b2c6-3b81a1cd79af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 15:02:52 crc kubenswrapper[4788]: E1010 15:02:52.999308 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" podUID="f3166185-7572-4637-b2c6-3b81a1cd79af" Oct 10 15:02:53 crc kubenswrapper[4788]: E1010 15:02:53.000329 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" podUID="d24a6fbc-3938-4612-a928-c0d97cba7fa8" Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.471400 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" event={"ID":"f3166185-7572-4637-b2c6-3b81a1cd79af","Type":"ContainerStarted","Data":"ca226317446ea631f55f32b39b9431f3cf7d3bb5357a7f6739c36abc0b047ba8"} Oct 10 15:02:53 crc kubenswrapper[4788]: E1010 15:02:53.473781 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" podUID="f3166185-7572-4637-b2c6-3b81a1cd79af" Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.474952 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" event={"ID":"84cbe7f0-4f13-468c-af71-022960cd0973","Type":"ContainerStarted","Data":"86508bc79112ea80563f9c033ec6bee12aa6b298461b821bbb3062a2e9d74b7c"} Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.476893 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" event={"ID":"9c8f6b95-2440-4fd6-b0a9-da1b17facf7e","Type":"ContainerStarted","Data":"a9f2b581c9da7d590acff09e10a54a001b7facaa13e05cf5c0227c6759948c1a"} Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.494066 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" event={"ID":"07597e46-5cc8-495d-95a1-f94235f2f9ed","Type":"ContainerStarted","Data":"1f9f25709feb3ff3c4f1bc4119e805c0b5ee19c77d6414ba67ad829804d705a4"} Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.496568 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" event={"ID":"d24a6fbc-3938-4612-a928-c0d97cba7fa8","Type":"ContainerStarted","Data":"985b63b198265b8d3c7f24725def9d3dd70ffc601195ef2914327c033a8b093e"} Oct 10 15:02:53 crc kubenswrapper[4788]: E1010 15:02:53.526035 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" podUID="d24a6fbc-3938-4612-a928-c0d97cba7fa8" Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.529924 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" event={"ID":"aae7a82d-e804-4f31-a8be-cc52317be3f1","Type":"ContainerStarted","Data":"2fc38edd9163a54d072ad3ca0a2d79b43736f7df60f02bc65995db86000e6396"} Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.574851 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" event={"ID":"2f9d641b-50d5-4dd9-baec-3c22d23273e1","Type":"ContainerStarted","Data":"38c777f73ade753664c60a3e2ccf9da2be75f09cdfe32c8c8ab8283edd043f52"} Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.578657 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv" event={"ID":"7940c932-25d0-4f02-bcd5-e3441aa026c4","Type":"ContainerStarted","Data":"19f1a29513fb9d9e48c2256486cd32a42e45458b0b8ef41d429034fd0f0ce5a2"} Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.586381 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" event={"ID":"22f63d73-f623-4b6a-83d7-dff55231ecc7","Type":"ContainerStarted","Data":"d07ae648819485613c2992d365af698a583e4af46df836c37f065efc32c10337"} Oct 10 15:02:53 crc kubenswrapper[4788]: I1010 15:02:53.602928 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv" podStartSLOduration=2.868419292 podStartE2EDuration="15.602899605s" podCreationTimestamp="2025-10-10 15:02:38 +0000 UTC" firstStartedPulling="2025-10-10 15:02:40.024553698 +0000 UTC m=+1062.474269246" lastFinishedPulling="2025-10-10 15:02:52.759034011 +0000 UTC m=+1075.208749559" observedRunningTime="2025-10-10 15:02:53.600602423 +0000 UTC m=+1076.050317961" watchObservedRunningTime="2025-10-10 15:02:53.602899605 +0000 UTC m=+1076.052615143" Oct 10 15:02:54 crc kubenswrapper[4788]: I1010 15:02:54.598958 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" event={"ID":"8d40607c-b865-4722-93db-94f3fb0b2f42","Type":"ContainerStarted","Data":"717bbacd983ca98e5089d134843a8e01150dd5ac05f88ae82c51d1cd9b1121ae"} Oct 10 15:02:54 crc kubenswrapper[4788]: I1010 15:02:54.600688 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" event={"ID":"38f6b32d-a5f6-402f-8070-eb1252c7d669","Type":"ContainerStarted","Data":"c453011b7749e6b1eb595262d21b4addd2c83eb92e96e6334ccf01cdff12a5b7"} Oct 10 15:02:54 crc kubenswrapper[4788]: I1010 15:02:54.603134 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" event={"ID":"1d47d678-34c1-41d7-a8cf-6f13199cc741","Type":"ContainerStarted","Data":"28da8b467d570c03e744ca06e6248e11531f8d424d436b777967792e8824f332"} Oct 10 15:02:54 crc kubenswrapper[4788]: I1010 15:02:54.606319 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" event={"ID":"45ccc74a-b207-4512-86fd-07e0495d1378","Type":"ContainerStarted","Data":"7c323fc1dbfadaac2e239d0e352ab097ef4fad177dd94f123edce17bad9fb984"} Oct 10 15:02:54 crc kubenswrapper[4788]: I1010 15:02:54.610819 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" event={"ID":"6791183a-bdf7-4ce2-b888-cd7e1b70ab80","Type":"ContainerStarted","Data":"616b8407a7e5f24f04e74feabfa4833fad740fdba97285f52d9ba2f7a950fc97"} Oct 10 15:02:54 crc kubenswrapper[4788]: E1010 15:02:54.613051 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" podUID="d24a6fbc-3938-4612-a928-c0d97cba7fa8" Oct 10 15:02:54 crc kubenswrapper[4788]: E1010 15:02:54.614053 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" podUID="f3166185-7572-4637-b2c6-3b81a1cd79af" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.652090 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" event={"ID":"38f6b32d-a5f6-402f-8070-eb1252c7d669","Type":"ContainerStarted","Data":"ee391cc2012b7baf2973a8af3bc1024b8f00a9dbc542a645183afc37d9d8059a"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.652567 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.655919 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.663745 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" event={"ID":"aae7a82d-e804-4f31-a8be-cc52317be3f1","Type":"ContainerStarted","Data":"4ed01bd51c85a36c77220a1bb407915e24d82dc65aae55280ac9662b59285b02"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.663948 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.666199 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.666649 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" event={"ID":"ebfc2c03-b35c-41a7-9679-11677bca30b2","Type":"ContainerStarted","Data":"336bb2c9778dbb4457cf16f5e057a846c37afeac62ec711206621fcc05ec4110"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.666680 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" event={"ID":"ebfc2c03-b35c-41a7-9679-11677bca30b2","Type":"ContainerStarted","Data":"cf9f001b76c0a7e121329b5c2210ff1e2639a89e8ee970a06336e87996c26f56"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.667062 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.670364 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-2z2xl" podStartSLOduration=7.61416784 podStartE2EDuration="20.670340074s" podCreationTimestamp="2025-10-10 15:02:38 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.765286815 +0000 UTC m=+1062.215002363" lastFinishedPulling="2025-10-10 15:02:52.821459039 +0000 UTC m=+1075.271174597" observedRunningTime="2025-10-10 15:02:58.669676676 +0000 UTC m=+1081.119392224" watchObservedRunningTime="2025-10-10 15:02:58.670340074 +0000 UTC m=+1081.120055622" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.680390 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" event={"ID":"8d40607c-b865-4722-93db-94f3fb0b2f42","Type":"ContainerStarted","Data":"4e7500183a19a5461f77bf4409df6572858e4030b2a0a345f0554e175024d255"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.680596 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.683383 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.686962 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" event={"ID":"22f63d73-f623-4b6a-83d7-dff55231ecc7","Type":"ContainerStarted","Data":"a8c6c7de3f30835fffafa8394b02f4f696545e75cea1a897832582ff21982ed0"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.687507 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.689095 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.690371 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" event={"ID":"45ccc74a-b207-4512-86fd-07e0495d1378","Type":"ContainerStarted","Data":"2b1a7f29d1d5680a8449777cfda6ef41e9cac8d78dd9c564d2e72f96c2347cd9"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.690823 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.696836 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" event={"ID":"2f9d641b-50d5-4dd9-baec-3c22d23273e1","Type":"ContainerStarted","Data":"b12a67e95057b905d17a5d7d509e956deb8af0a39c1d376a8f7e937e63c3958a"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.697898 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.703415 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.710293 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" podStartSLOduration=8.110187976 podStartE2EDuration="20.710272774s" podCreationTimestamp="2025-10-10 15:02:38 +0000 UTC" firstStartedPulling="2025-10-10 15:02:40.209049018 +0000 UTC m=+1062.658764566" lastFinishedPulling="2025-10-10 15:02:52.809133816 +0000 UTC m=+1075.258849364" observedRunningTime="2025-10-10 15:02:58.695234707 +0000 UTC m=+1081.144950265" watchObservedRunningTime="2025-10-10 15:02:58.710272774 +0000 UTC m=+1081.159988322" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.711003 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.712106 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" event={"ID":"9c8f6b95-2440-4fd6-b0a9-da1b17facf7e","Type":"ContainerStarted","Data":"0082e67d4b9682d5586c180463dfe8d9974f006e32291445537fef7ddda7813d"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.713028 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.716648 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.726374 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" event={"ID":"1d47d678-34c1-41d7-a8cf-6f13199cc741","Type":"ContainerStarted","Data":"619e7aa5ae53c0e850bd0f76f31863b23f0b6cd636fb092bf4f04f9a3f562f50"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.728517 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.729178 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.764422 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" event={"ID":"41d202a5-eddf-4ee7-9b9a-9edf7905d51f","Type":"ContainerStarted","Data":"06f1284d0a0f726018b645e89e62011e4cf254fc36d2a703cb4504baf17970da"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.764475 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" event={"ID":"41d202a5-eddf-4ee7-9b9a-9edf7905d51f","Type":"ContainerStarted","Data":"0d4f497581b476758c1d7f38b3a0db463dbc348d43cd2d20ca519e524d39a567"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.765197 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.788551 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" event={"ID":"67694f95-9859-40b3-ae66-65cfdc1d17f5","Type":"ContainerStarted","Data":"4d7aa91f28237d6947f7c64fc005331ecb16236d967dd69cd3413313ef72b459"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.792049 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.792298 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" event={"ID":"67694f95-9859-40b3-ae66-65cfdc1d17f5","Type":"ContainerStarted","Data":"d0868eb1ae1cbed20ce95e8bc815a070d73f3e3c70e6c61f48eeff98465603dc"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.792387 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" event={"ID":"84cbe7f0-4f13-468c-af71-022960cd0973","Type":"ContainerStarted","Data":"868d13187a0d246947461cea9f4d087c643517a5448c7b2e9693d3597372865a"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.788647 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-5v88l" podStartSLOduration=8.564526398 podStartE2EDuration="21.788628824s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.472294591 +0000 UTC m=+1061.922010139" lastFinishedPulling="2025-10-10 15:02:52.696397017 +0000 UTC m=+1075.146112565" observedRunningTime="2025-10-10 15:02:58.788199892 +0000 UTC m=+1081.237915440" watchObservedRunningTime="2025-10-10 15:02:58.788628824 +0000 UTC m=+1081.238344372" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.793813 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.798216 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" event={"ID":"6791183a-bdf7-4ce2-b888-cd7e1b70ab80","Type":"ContainerStarted","Data":"da4b9e1dd21accd2e7a56afa7ff086358d7493fbf62406d150bd0b5e92e99f25"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.800032 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.803284 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-8l7ct" podStartSLOduration=8.349197364 podStartE2EDuration="21.803253819s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.304701808 +0000 UTC m=+1061.754417346" lastFinishedPulling="2025-10-10 15:02:52.758758253 +0000 UTC m=+1075.208473801" observedRunningTime="2025-10-10 15:02:58.749048003 +0000 UTC m=+1081.198763551" watchObservedRunningTime="2025-10-10 15:02:58.803253819 +0000 UTC m=+1081.252969377" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.813889 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.816498 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.818464 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" event={"ID":"07597e46-5cc8-495d-95a1-f94235f2f9ed","Type":"ContainerStarted","Data":"9c09d98eb5b113cd14893e56c5dc299500ad7cef120a84eb917cba33ccc3c875"} Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.819626 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.854859 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.893742 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-j4xpv" podStartSLOduration=8.882316123 podStartE2EDuration="21.893716596s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.808592207 +0000 UTC m=+1062.258307755" lastFinishedPulling="2025-10-10 15:02:52.81999269 +0000 UTC m=+1075.269708228" observedRunningTime="2025-10-10 15:02:58.847105795 +0000 UTC m=+1081.296821343" watchObservedRunningTime="2025-10-10 15:02:58.893716596 +0000 UTC m=+1081.343432144" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.897331 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-jvfjk" podStartSLOduration=7.572925286 podStartE2EDuration="20.897323674s" podCreationTimestamp="2025-10-10 15:02:38 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.473131294 +0000 UTC m=+1061.922846842" lastFinishedPulling="2025-10-10 15:02:52.797529682 +0000 UTC m=+1075.247245230" observedRunningTime="2025-10-10 15:02:58.892597345 +0000 UTC m=+1081.342312893" watchObservedRunningTime="2025-10-10 15:02:58.897323674 +0000 UTC m=+1081.347039222" Oct 10 15:02:58 crc kubenswrapper[4788]: I1010 15:02:58.962433 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-k5ls7" podStartSLOduration=8.169019091 podStartE2EDuration="21.962409104s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:38.987308954 +0000 UTC m=+1061.437024502" lastFinishedPulling="2025-10-10 15:02:52.780698967 +0000 UTC m=+1075.230414515" observedRunningTime="2025-10-10 15:02:58.952494346 +0000 UTC m=+1081.402209894" watchObservedRunningTime="2025-10-10 15:02:58.962409104 +0000 UTC m=+1081.412124652" Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.004658 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-929g7" podStartSLOduration=8.31985041 podStartE2EDuration="22.004636566s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.00751637 +0000 UTC m=+1061.457231918" lastFinishedPulling="2025-10-10 15:02:52.692302526 +0000 UTC m=+1075.142018074" observedRunningTime="2025-10-10 15:02:59.002988081 +0000 UTC m=+1081.452703619" watchObservedRunningTime="2025-10-10 15:02:59.004636566 +0000 UTC m=+1081.454352114" Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.064059 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-9pbsq" podStartSLOduration=8.496059857 podStartE2EDuration="22.064028433s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.229470874 +0000 UTC m=+1061.679186422" lastFinishedPulling="2025-10-10 15:02:52.79743945 +0000 UTC m=+1075.247154998" observedRunningTime="2025-10-10 15:02:59.053770174 +0000 UTC m=+1081.503485722" watchObservedRunningTime="2025-10-10 15:02:59.064028433 +0000 UTC m=+1081.513743981" Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.118431 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz" podStartSLOduration=9.950480234 podStartE2EDuration="22.118409853s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:40.612952383 +0000 UTC m=+1063.062667941" lastFinishedPulling="2025-10-10 15:02:52.780882012 +0000 UTC m=+1075.230597560" observedRunningTime="2025-10-10 15:02:59.111801715 +0000 UTC m=+1081.561517263" watchObservedRunningTime="2025-10-10 15:02:59.118409853 +0000 UTC m=+1081.568125541" Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.151128 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" podStartSLOduration=8.674491671 podStartE2EDuration="22.151090447s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.282400495 +0000 UTC m=+1061.732116043" lastFinishedPulling="2025-10-10 15:02:52.758999271 +0000 UTC m=+1075.208714819" observedRunningTime="2025-10-10 15:02:59.138812345 +0000 UTC m=+1081.588527893" watchObservedRunningTime="2025-10-10 15:02:59.151090447 +0000 UTC m=+1081.600805995" Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.184518 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" podStartSLOduration=9.193654505 podStartE2EDuration="22.184492431s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.819909943 +0000 UTC m=+1062.269625491" lastFinishedPulling="2025-10-10 15:02:52.810747869 +0000 UTC m=+1075.260463417" observedRunningTime="2025-10-10 15:02:59.182473646 +0000 UTC m=+1081.632189194" watchObservedRunningTime="2025-10-10 15:02:59.184492431 +0000 UTC m=+1081.634207979" Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.211964 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ckw79" podStartSLOduration=8.522581343 podStartE2EDuration="22.211937633s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.090503574 +0000 UTC m=+1061.540219122" lastFinishedPulling="2025-10-10 15:02:52.779859864 +0000 UTC m=+1075.229575412" observedRunningTime="2025-10-10 15:02:59.207594025 +0000 UTC m=+1081.657309573" watchObservedRunningTime="2025-10-10 15:02:59.211937633 +0000 UTC m=+1081.661653181" Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.242530 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-x2mn7" podStartSLOduration=9.243237534 podStartE2EDuration="22.2425083s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.796117819 +0000 UTC m=+1062.245833367" lastFinishedPulling="2025-10-10 15:02:52.795388575 +0000 UTC m=+1075.245104133" observedRunningTime="2025-10-10 15:02:59.236658581 +0000 UTC m=+1081.686374129" watchObservedRunningTime="2025-10-10 15:02:59.2425083 +0000 UTC m=+1081.692223848" Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.405756 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:02:59 crc kubenswrapper[4788]: I1010 15:02:59.405818 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.862045 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" event={"ID":"c2dbece7-c07a-4223-98a2-bb96f93b9e3c","Type":"ContainerStarted","Data":"f25cfedeafbc7404520d25f07e6185488a532817a657f52e492f824b01fba4a1"} Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.862725 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.864826 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" event={"ID":"682f41df-b648-476f-8a3e-9475501bf290","Type":"ContainerStarted","Data":"bc45d7caa430da2961ad0eba12e70abd1b612e7b4944efde2ea5a38e40a25739"} Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.865432 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.883271 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" event={"ID":"d99a1777-7ea0-4072-b086-aabc93f57f71","Type":"ContainerStarted","Data":"bbf56d0e1bf032c1903c905cbb26a83022a090177657ee027da3222a55a566b3"} Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.883440 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" event={"ID":"4e656fbd-3ed0-4434-9264-5a08a20607aa","Type":"ContainerStarted","Data":"02a5f4fc40e55ccae7195da138e0747eb49b35ecd4af80a6e09beb0d00c9445c"} Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.883467 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" event={"ID":"6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0","Type":"ContainerStarted","Data":"2fff22f03d7e179cea02a3a42a92d7d2bb4e5ea8e31c9eae426309c231a9ea84"} Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.885307 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.886193 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.886476 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.938788 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" podStartSLOduration=2.620520517 podStartE2EDuration="24.938762523s" podCreationTimestamp="2025-10-10 15:02:38 +0000 UTC" firstStartedPulling="2025-10-10 15:02:40.076216815 +0000 UTC m=+1062.525932353" lastFinishedPulling="2025-10-10 15:03:02.394458801 +0000 UTC m=+1084.844174359" observedRunningTime="2025-10-10 15:03:02.898000661 +0000 UTC m=+1085.347716209" watchObservedRunningTime="2025-10-10 15:03:02.938762523 +0000 UTC m=+1085.388478071" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.962519 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" podStartSLOduration=3.600470795 podStartE2EDuration="25.962497865s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:40.039451001 +0000 UTC m=+1062.489166549" lastFinishedPulling="2025-10-10 15:03:02.401478081 +0000 UTC m=+1084.851193619" observedRunningTime="2025-10-10 15:03:02.919960835 +0000 UTC m=+1085.369676383" watchObservedRunningTime="2025-10-10 15:03:02.962497865 +0000 UTC m=+1085.412213403" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.980724 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" podStartSLOduration=3.68458518 podStartE2EDuration="25.980702247s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:40.093646357 +0000 UTC m=+1062.543361905" lastFinishedPulling="2025-10-10 15:03:02.389763424 +0000 UTC m=+1084.839478972" observedRunningTime="2025-10-10 15:03:02.949696239 +0000 UTC m=+1085.399411807" watchObservedRunningTime="2025-10-10 15:03:02.980702247 +0000 UTC m=+1085.430417795" Oct 10 15:03:02 crc kubenswrapper[4788]: I1010 15:03:02.982375 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" podStartSLOduration=3.6062294599999998 podStartE2EDuration="25.982368902s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:40.037299953 +0000 UTC m=+1062.487015501" lastFinishedPulling="2025-10-10 15:03:02.413439355 +0000 UTC m=+1084.863154943" observedRunningTime="2025-10-10 15:03:02.976786961 +0000 UTC m=+1085.426502509" watchObservedRunningTime="2025-10-10 15:03:02.982368902 +0000 UTC m=+1085.432084450" Oct 10 15:03:03 crc kubenswrapper[4788]: I1010 15:03:03.002614 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" podStartSLOduration=3.6599864760000003 podStartE2EDuration="26.00259507s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:40.047234612 +0000 UTC m=+1062.496950160" lastFinishedPulling="2025-10-10 15:03:02.389843216 +0000 UTC m=+1084.839558754" observedRunningTime="2025-10-10 15:03:03.000616246 +0000 UTC m=+1085.450331804" watchObservedRunningTime="2025-10-10 15:03:03.00259507 +0000 UTC m=+1085.452310618" Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.241365 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.317705 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-4spql" Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.440997 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-pkc5n" Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.499079 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-4gq6b" Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.559329 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-mvw9d" Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.594290 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-2cwz7" Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.601515 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-646675d848-hcq88" Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.871308 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-7tlj8" Oct 10 15:03:08 crc kubenswrapper[4788]: I1010 15:03:08.874269 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-l678z" Oct 10 15:03:09 crc kubenswrapper[4788]: I1010 15:03:09.952932 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" event={"ID":"d24a6fbc-3938-4612-a928-c0d97cba7fa8","Type":"ContainerStarted","Data":"68fd7da94160daa5199aad820ee8d525e705f6d8012717869f36fda47ca726e9"} Oct 10 15:03:09 crc kubenswrapper[4788]: I1010 15:03:09.953417 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" Oct 10 15:03:09 crc kubenswrapper[4788]: I1010 15:03:09.985098 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" podStartSLOduration=3.502815624 podStartE2EDuration="32.985065926s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.45562667 +0000 UTC m=+1061.905342218" lastFinishedPulling="2025-10-10 15:03:08.937876972 +0000 UTC m=+1091.387592520" observedRunningTime="2025-10-10 15:03:09.97561331 +0000 UTC m=+1092.425328858" watchObservedRunningTime="2025-10-10 15:03:09.985065926 +0000 UTC m=+1092.434781514" Oct 10 15:03:10 crc kubenswrapper[4788]: I1010 15:03:10.968832 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" event={"ID":"f3166185-7572-4637-b2c6-3b81a1cd79af","Type":"ContainerStarted","Data":"e613ca3ecbfdc55588488fce9c6fd48acb0092f2b7116f874fae10fa20b299b6"} Oct 10 15:03:10 crc kubenswrapper[4788]: I1010 15:03:10.969746 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" Oct 10 15:03:10 crc kubenswrapper[4788]: I1010 15:03:10.997078 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" podStartSLOduration=2.559692256 podStartE2EDuration="33.997052457s" podCreationTimestamp="2025-10-10 15:02:37 +0000 UTC" firstStartedPulling="2025-10-10 15:02:39.229417772 +0000 UTC m=+1061.679133320" lastFinishedPulling="2025-10-10 15:03:10.666777973 +0000 UTC m=+1093.116493521" observedRunningTime="2025-10-10 15:03:10.991816306 +0000 UTC m=+1093.441531864" watchObservedRunningTime="2025-10-10 15:03:10.997052457 +0000 UTC m=+1093.446768015" Oct 10 15:03:18 crc kubenswrapper[4788]: I1010 15:03:18.412838 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-5hstt" Oct 10 15:03:18 crc kubenswrapper[4788]: I1010 15:03:18.465029 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kbw8c" Oct 10 15:03:29 crc kubenswrapper[4788]: I1010 15:03:29.406754 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:03:29 crc kubenswrapper[4788]: I1010 15:03:29.407496 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:03:29 crc kubenswrapper[4788]: I1010 15:03:29.407549 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:03:29 crc kubenswrapper[4788]: I1010 15:03:29.408132 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bf2bdef66313687037ed74946bc83398b51a4a5e0533a877fae35897091d486"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:03:29 crc kubenswrapper[4788]: I1010 15:03:29.408224 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://9bf2bdef66313687037ed74946bc83398b51a4a5e0533a877fae35897091d486" gracePeriod=600 Oct 10 15:03:30 crc kubenswrapper[4788]: I1010 15:03:30.119515 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="9bf2bdef66313687037ed74946bc83398b51a4a5e0533a877fae35897091d486" exitCode=0 Oct 10 15:03:30 crc kubenswrapper[4788]: I1010 15:03:30.119599 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"9bf2bdef66313687037ed74946bc83398b51a4a5e0533a877fae35897091d486"} Oct 10 15:03:30 crc kubenswrapper[4788]: I1010 15:03:30.120257 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"8e260922adcdc34acddb0d03ca667c174167bee6681d51128f1688ce0f11d52b"} Oct 10 15:03:30 crc kubenswrapper[4788]: I1010 15:03:30.120309 4788 scope.go:117] "RemoveContainer" containerID="d62f45676977d2854ac3c581616b717de6d75ef758d15d869e78e7d7a3a145f9" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.850853 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6lc8q"] Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.852597 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.860410 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.860938 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.860950 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.861263 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-szb4h" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.866755 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6lc8q"] Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.903047 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxf7t\" (UniqueName: \"kubernetes.io/projected/b309e015-8f4c-4c29-bf9c-7aa1e822c721-kube-api-access-vxf7t\") pod \"dnsmasq-dns-675f4bcbfc-6lc8q\" (UID: \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.903111 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b309e015-8f4c-4c29-bf9c-7aa1e822c721-config\") pod \"dnsmasq-dns-675f4bcbfc-6lc8q\" (UID: \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.925380 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vfvzm"] Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.927663 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.932568 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 10 15:03:35 crc kubenswrapper[4788]: I1010 15:03:35.941067 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vfvzm"] Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.003898 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.003965 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7x7x\" (UniqueName: \"kubernetes.io/projected/c4e363ab-5f3f-45a6-8840-03bc08acacba-kube-api-access-g7x7x\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.004007 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxf7t\" (UniqueName: \"kubernetes.io/projected/b309e015-8f4c-4c29-bf9c-7aa1e822c721-kube-api-access-vxf7t\") pod \"dnsmasq-dns-675f4bcbfc-6lc8q\" (UID: \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.004033 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b309e015-8f4c-4c29-bf9c-7aa1e822c721-config\") pod \"dnsmasq-dns-675f4bcbfc-6lc8q\" (UID: \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.004047 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-config\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.005165 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b309e015-8f4c-4c29-bf9c-7aa1e822c721-config\") pod \"dnsmasq-dns-675f4bcbfc-6lc8q\" (UID: \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.053195 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxf7t\" (UniqueName: \"kubernetes.io/projected/b309e015-8f4c-4c29-bf9c-7aa1e822c721-kube-api-access-vxf7t\") pod \"dnsmasq-dns-675f4bcbfc-6lc8q\" (UID: \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.104481 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.104560 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7x7x\" (UniqueName: \"kubernetes.io/projected/c4e363ab-5f3f-45a6-8840-03bc08acacba-kube-api-access-g7x7x\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.104612 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-config\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.105423 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-config\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.105424 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.122031 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7x7x\" (UniqueName: \"kubernetes.io/projected/c4e363ab-5f3f-45a6-8840-03bc08acacba-kube-api-access-g7x7x\") pod \"dnsmasq-dns-78dd6ddcc-vfvzm\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.177006 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.247607 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.676079 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6lc8q"] Oct 10 15:03:36 crc kubenswrapper[4788]: I1010 15:03:36.754653 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vfvzm"] Oct 10 15:03:37 crc kubenswrapper[4788]: I1010 15:03:37.176161 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" event={"ID":"c4e363ab-5f3f-45a6-8840-03bc08acacba","Type":"ContainerStarted","Data":"c8a4fddf10e37ec4cfc982eb9aa34a704cdc874dcad23e180c9350f5fd822e8c"} Oct 10 15:03:37 crc kubenswrapper[4788]: I1010 15:03:37.177316 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" event={"ID":"b309e015-8f4c-4c29-bf9c-7aa1e822c721","Type":"ContainerStarted","Data":"2b61c5bcd4a8e5adace3fe8819de2c21f1776afa2f8d9aa98f3feb6091fd7ea5"} Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.716732 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6lc8q"] Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.758957 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jfxp2"] Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.760381 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.781409 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jfxp2"] Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.873779 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-config\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.873861 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.873887 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r25nd\" (UniqueName: \"kubernetes.io/projected/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-kube-api-access-r25nd\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.977824 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-config\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.977911 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.977939 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r25nd\" (UniqueName: \"kubernetes.io/projected/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-kube-api-access-r25nd\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.979239 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-config\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:38 crc kubenswrapper[4788]: I1010 15:03:38.980117 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.008391 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r25nd\" (UniqueName: \"kubernetes.io/projected/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-kube-api-access-r25nd\") pod \"dnsmasq-dns-666b6646f7-jfxp2\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.073374 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vfvzm"] Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.099261 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-r4mvh"] Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.101178 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.113876 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-r4mvh"] Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.124499 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.189865 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dqvc\" (UniqueName: \"kubernetes.io/projected/9e62dafe-2ab3-467f-adda-d18766ddd450-kube-api-access-6dqvc\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.189977 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.190038 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-config\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.292036 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.292248 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-config\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.292310 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dqvc\" (UniqueName: \"kubernetes.io/projected/9e62dafe-2ab3-467f-adda-d18766ddd450-kube-api-access-6dqvc\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.294066 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.295404 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-config\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.317817 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dqvc\" (UniqueName: \"kubernetes.io/projected/9e62dafe-2ab3-467f-adda-d18766ddd450-kube-api-access-6dqvc\") pod \"dnsmasq-dns-57d769cc4f-r4mvh\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.444330 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.687573 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jfxp2"] Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.932438 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.933679 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.936640 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.937353 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.938384 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.938551 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.938637 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.942455 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.942473 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-z7cdn" Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.976998 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 15:03:39 crc kubenswrapper[4788]: I1010 15:03:39.986520 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-r4mvh"] Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.105771 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dcf4b178-75ee-47fb-81e4-e6b277eb1790-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106348 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106408 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106441 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106474 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106506 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x97t\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-kube-api-access-2x97t\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106552 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106585 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dcf4b178-75ee-47fb-81e4-e6b277eb1790-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106662 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106696 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.106723 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.207921 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x97t\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-kube-api-access-2x97t\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208008 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208048 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dcf4b178-75ee-47fb-81e4-e6b277eb1790-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208114 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208161 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208181 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208206 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dcf4b178-75ee-47fb-81e4-e6b277eb1790-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208234 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208269 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208290 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.208314 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.209413 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.209732 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.209839 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.210264 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.210684 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.212453 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.216405 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.216684 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dcf4b178-75ee-47fb-81e4-e6b277eb1790-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.216967 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.229324 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dcf4b178-75ee-47fb-81e4-e6b277eb1790-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.281988 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x97t\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-kube-api-access-2x97t\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.325845 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" event={"ID":"9e62dafe-2ab3-467f-adda-d18766ddd450","Type":"ContainerStarted","Data":"314fad5b5d8e18c796b7751ac1bc02f583c7e0e0acf335937275d9ff287e674a"} Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.332982 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" event={"ID":"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a","Type":"ContainerStarted","Data":"3ebb16c38bd727fba2bab9c66dea87220408b0f4286ab515910f447d62cc8bef"} Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.333976 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.335866 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.344231 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.345363 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.345597 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.345730 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.345945 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-jmj4b" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.352915 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.353214 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.353361 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.354324 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.420974 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421065 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421126 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421288 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421360 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421377 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421399 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421459 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421479 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421572 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.421594 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z55k\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-kube-api-access-8z55k\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524599 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524673 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524692 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524716 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524758 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524779 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524855 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524921 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524949 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z55k\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-kube-api-access-8z55k\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524971 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.524990 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.525011 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.525669 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.526603 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.527769 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.527943 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.532861 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.534611 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.537103 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.538599 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.540022 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.556069 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z55k\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-kube-api-access-8z55k\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.561449 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.565827 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 15:03:40 crc kubenswrapper[4788]: I1010 15:03:40.693343 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.521156 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.523255 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.528657 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-lgjkm" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.528989 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.529441 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.529618 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.530045 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.535283 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.552732 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647185 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647246 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-kolla-config\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647272 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647317 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-secrets\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647346 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-default\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647365 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647392 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz4jq\" (UniqueName: \"kubernetes.io/projected/4c25a6d1-ee44-438c-a16d-e6952cfda420-kube-api-access-wz4jq\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647411 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.647431 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.748908 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-secrets\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.748977 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-default\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.749001 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.749028 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz4jq\" (UniqueName: \"kubernetes.io/projected/4c25a6d1-ee44-438c-a16d-e6952cfda420-kube-api-access-wz4jq\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.749050 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.749069 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.749113 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.749152 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-kolla-config\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.749175 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.753981 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.754345 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-default\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.755768 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.755962 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-kolla-config\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.756950 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.765734 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.767955 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-secrets\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.771529 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.781206 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz4jq\" (UniqueName: \"kubernetes.io/projected/4c25a6d1-ee44-438c-a16d-e6952cfda420-kube-api-access-wz4jq\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.794330 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " pod="openstack/openstack-galera-0" Oct 10 15:03:41 crc kubenswrapper[4788]: I1010 15:03:41.865299 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.798124 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.799606 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.802365 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.803088 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8dnqq" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.803233 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.803527 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.813837 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.879196 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.879383 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.879636 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.879758 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.879812 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.879995 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.880016 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cwh4\" (UniqueName: \"kubernetes.io/projected/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kube-api-access-8cwh4\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.880060 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.880127 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981385 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981433 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cwh4\" (UniqueName: \"kubernetes.io/projected/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kube-api-access-8cwh4\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981457 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981484 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981506 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981529 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981578 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981610 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.981629 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.982045 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.982586 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.983737 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.984513 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.985328 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.992006 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:42 crc kubenswrapper[4788]: I1010 15:03:42.992387 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.007398 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.007421 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cwh4\" (UniqueName: \"kubernetes.io/projected/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kube-api-access-8cwh4\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.043084 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.140921 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.198976 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.200477 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.204119 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.204574 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-6cz95" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.207551 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.227912 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.292484 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgk4f\" (UniqueName: \"kubernetes.io/projected/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kube-api-access-zgk4f\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.292556 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kolla-config\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.292666 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-config-data\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.293623 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.293699 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.395361 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.395456 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.395515 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgk4f\" (UniqueName: \"kubernetes.io/projected/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kube-api-access-zgk4f\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.395534 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kolla-config\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.395577 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-config-data\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.396934 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-config-data\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.398326 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kolla-config\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.403254 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.403716 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.423215 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgk4f\" (UniqueName: \"kubernetes.io/projected/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kube-api-access-zgk4f\") pod \"memcached-0\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " pod="openstack/memcached-0" Oct 10 15:03:43 crc kubenswrapper[4788]: I1010 15:03:43.531015 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 15:03:45 crc kubenswrapper[4788]: I1010 15:03:45.076198 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:03:45 crc kubenswrapper[4788]: I1010 15:03:45.079090 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 15:03:45 crc kubenswrapper[4788]: I1010 15:03:45.085167 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-m52sc" Oct 10 15:03:45 crc kubenswrapper[4788]: I1010 15:03:45.095329 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:03:45 crc kubenswrapper[4788]: I1010 15:03:45.132244 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trkbn\" (UniqueName: \"kubernetes.io/projected/23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c-kube-api-access-trkbn\") pod \"kube-state-metrics-0\" (UID: \"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c\") " pod="openstack/kube-state-metrics-0" Oct 10 15:03:45 crc kubenswrapper[4788]: I1010 15:03:45.234659 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trkbn\" (UniqueName: \"kubernetes.io/projected/23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c-kube-api-access-trkbn\") pod \"kube-state-metrics-0\" (UID: \"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c\") " pod="openstack/kube-state-metrics-0" Oct 10 15:03:45 crc kubenswrapper[4788]: I1010 15:03:45.256472 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trkbn\" (UniqueName: \"kubernetes.io/projected/23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c-kube-api-access-trkbn\") pod \"kube-state-metrics-0\" (UID: \"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c\") " pod="openstack/kube-state-metrics-0" Oct 10 15:03:45 crc kubenswrapper[4788]: I1010 15:03:45.409327 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.457639 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-n84gv"] Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.459877 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.462841 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.463206 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.463345 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-jdsst" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.470102 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-n84gv"] Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.509679 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.509749 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-scripts\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.509776 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-ovn-controller-tls-certs\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.509814 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjk25\" (UniqueName: \"kubernetes.io/projected/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-kube-api-access-hjk25\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.509876 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run-ovn\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.509894 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-log-ovn\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.509910 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-combined-ca-bundle\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.517015 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-wpgvn"] Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.520461 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.540120 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-wpgvn"] Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.611714 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7jns\" (UniqueName: \"kubernetes.io/projected/6627a65b-765d-430c-99f5-eddd285d4289-kube-api-access-n7jns\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.611765 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6627a65b-765d-430c-99f5-eddd285d4289-scripts\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.611861 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.611895 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-scripts\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.611920 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-ovn-controller-tls-certs\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.611947 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjk25\" (UniqueName: \"kubernetes.io/projected/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-kube-api-access-hjk25\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.611976 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-log\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.612035 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-run\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.612057 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run-ovn\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.612084 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-log-ovn\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.612112 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-lib\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.612181 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-combined-ca-bundle\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.612208 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-etc-ovs\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.612939 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.613874 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-log-ovn\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.614046 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run-ovn\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.615615 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-scripts\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.618731 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-combined-ca-bundle\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.619688 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-ovn-controller-tls-certs\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.632670 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjk25\" (UniqueName: \"kubernetes.io/projected/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-kube-api-access-hjk25\") pod \"ovn-controller-n84gv\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.713930 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6627a65b-765d-430c-99f5-eddd285d4289-scripts\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714398 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7jns\" (UniqueName: \"kubernetes.io/projected/6627a65b-765d-430c-99f5-eddd285d4289-kube-api-access-n7jns\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714477 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-log\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714543 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-run\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714574 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-lib\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714606 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-etc-ovs\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714754 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-run\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714883 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-etc-ovs\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714928 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-log\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.714957 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-lib\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.716280 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6627a65b-765d-430c-99f5-eddd285d4289-scripts\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.731980 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7jns\" (UniqueName: \"kubernetes.io/projected/6627a65b-765d-430c-99f5-eddd285d4289-kube-api-access-n7jns\") pod \"ovn-controller-ovs-wpgvn\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.786026 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n84gv" Oct 10 15:03:48 crc kubenswrapper[4788]: I1010 15:03:48.849657 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.235160 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.237739 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.241677 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.241669 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.241949 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-pflzz" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.242088 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.242235 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.250782 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.327025 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.327091 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.327121 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.327160 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km8v9\" (UniqueName: \"kubernetes.io/projected/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-kube-api-access-km8v9\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.327225 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.327277 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.327296 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.327339 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-config\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.429496 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.429556 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.429599 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-config\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.429667 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.429700 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.429726 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.429748 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km8v9\" (UniqueName: \"kubernetes.io/projected/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-kube-api-access-km8v9\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.429791 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.433797 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.434345 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-config\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.436727 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.437784 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.437821 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.445295 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.450273 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.459171 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km8v9\" (UniqueName: \"kubernetes.io/projected/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-kube-api-access-km8v9\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.467130 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:49 crc kubenswrapper[4788]: I1010 15:03:49.564620 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.371741 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.373942 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.377704 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.377893 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qx6kb" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.378120 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.378273 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.381609 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.505860 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.505953 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b927859-0238-46f5-90aa-4afa094c434e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.505982 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.506017 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.506042 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.506101 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.506119 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd69n\" (UniqueName: \"kubernetes.io/projected/4b927859-0238-46f5-90aa-4afa094c434e-kube-api-access-bd69n\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.506157 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.608173 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b927859-0238-46f5-90aa-4afa094c434e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.608216 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.608255 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.608281 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.608309 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.608326 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd69n\" (UniqueName: \"kubernetes.io/projected/4b927859-0238-46f5-90aa-4afa094c434e-kube-api-access-bd69n\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.608346 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.608382 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.609380 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.610452 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.611064 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-config\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.611402 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b927859-0238-46f5-90aa-4afa094c434e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.614697 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.621865 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.622536 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.630042 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd69n\" (UniqueName: \"kubernetes.io/projected/4b927859-0238-46f5-90aa-4afa094c434e-kube-api-access-bd69n\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.638212 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:52 crc kubenswrapper[4788]: I1010 15:03:52.714724 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 15:03:53 crc kubenswrapper[4788]: E1010 15:03:53.722014 4788 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 10 15:03:53 crc kubenswrapper[4788]: E1010 15:03:53.723002 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g7x7x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-vfvzm_openstack(c4e363ab-5f3f-45a6-8840-03bc08acacba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 15:03:53 crc kubenswrapper[4788]: E1010 15:03:53.724410 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" podUID="c4e363ab-5f3f-45a6-8840-03bc08acacba" Oct 10 15:03:53 crc kubenswrapper[4788]: E1010 15:03:53.778665 4788 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 10 15:03:53 crc kubenswrapper[4788]: E1010 15:03:53.779429 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vxf7t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-6lc8q_openstack(b309e015-8f4c-4c29-bf9c-7aa1e822c721): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 15:03:53 crc kubenswrapper[4788]: E1010 15:03:53.780895 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" podUID="b309e015-8f4c-4c29-bf9c-7aa1e822c721" Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.381956 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 10 15:03:54 crc kubenswrapper[4788]: W1010 15:03:54.384237 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4075bf4d_01ed_4d7c_9fb8_576dd2866821.slice/crio-1fc4538efddc96abbaea1530461b273bd133f7132bfb713b71d2632c7c770583 WatchSource:0}: Error finding container 1fc4538efddc96abbaea1530461b273bd133f7132bfb713b71d2632c7c770583: Status 404 returned error can't find the container with id 1fc4538efddc96abbaea1530461b273bd133f7132bfb713b71d2632c7c770583 Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.454515 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.462872 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.470992 4788 generic.go:334] "Generic (PLEG): container finished" podID="9e62dafe-2ab3-467f-adda-d18766ddd450" containerID="10d08519d4c7317c152b6029114795663f2138c64e9ebdd0c519113d68276b30" exitCode=0 Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.471069 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" event={"ID":"9e62dafe-2ab3-467f-adda-d18766ddd450","Type":"ContainerDied","Data":"10d08519d4c7317c152b6029114795663f2138c64e9ebdd0c519113d68276b30"} Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.479006 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.480460 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4075bf4d-01ed-4d7c-9fb8-576dd2866821","Type":"ContainerStarted","Data":"1fc4538efddc96abbaea1530461b273bd133f7132bfb713b71d2632c7c770583"} Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.483017 4788 generic.go:334] "Generic (PLEG): container finished" podID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" containerID="eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc" exitCode=0 Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.483327 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" event={"ID":"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a","Type":"ContainerDied","Data":"eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc"} Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.671586 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.691867 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-n84gv"] Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.699529 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.793974 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 15:03:54 crc kubenswrapper[4788]: W1010 15:03:54.803041 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8951c1b9_28fd_449d_ade9_a7b7bfc3af6e.slice/crio-9c05966ceccac74c34fc128399f37a6a0c27ea686de2c2efe9f4b3d85242332d WatchSource:0}: Error finding container 9c05966ceccac74c34fc128399f37a6a0c27ea686de2c2efe9f4b3d85242332d: Status 404 returned error can't find the container with id 9c05966ceccac74c34fc128399f37a6a0c27ea686de2c2efe9f4b3d85242332d Oct 10 15:03:54 crc kubenswrapper[4788]: E1010 15:03:54.897100 4788 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 10 15:03:54 crc kubenswrapper[4788]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 10 15:03:54 crc kubenswrapper[4788]: > podSandboxID="3ebb16c38bd727fba2bab9c66dea87220408b0f4286ab515910f447d62cc8bef" Oct 10 15:03:54 crc kubenswrapper[4788]: E1010 15:03:54.897350 4788 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 10 15:03:54 crc kubenswrapper[4788]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r25nd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-jfxp2_openstack(6bc95fb6-a85a-440d-8fcd-62386c6a3b5a): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 10 15:03:54 crc kubenswrapper[4788]: > logger="UnhandledError" Oct 10 15:03:54 crc kubenswrapper[4788]: E1010 15:03:54.899269 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" podUID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.986311 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:54 crc kubenswrapper[4788]: I1010 15:03:54.991054 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.162484 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b309e015-8f4c-4c29-bf9c-7aa1e822c721-config\") pod \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\" (UID: \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\") " Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.163226 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxf7t\" (UniqueName: \"kubernetes.io/projected/b309e015-8f4c-4c29-bf9c-7aa1e822c721-kube-api-access-vxf7t\") pod \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\" (UID: \"b309e015-8f4c-4c29-bf9c-7aa1e822c721\") " Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.163277 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b309e015-8f4c-4c29-bf9c-7aa1e822c721-config" (OuterVolumeSpecName: "config") pod "b309e015-8f4c-4c29-bf9c-7aa1e822c721" (UID: "b309e015-8f4c-4c29-bf9c-7aa1e822c721"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.163379 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-config\") pod \"c4e363ab-5f3f-45a6-8840-03bc08acacba\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.163415 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7x7x\" (UniqueName: \"kubernetes.io/projected/c4e363ab-5f3f-45a6-8840-03bc08acacba-kube-api-access-g7x7x\") pod \"c4e363ab-5f3f-45a6-8840-03bc08acacba\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.164113 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-config" (OuterVolumeSpecName: "config") pod "c4e363ab-5f3f-45a6-8840-03bc08acacba" (UID: "c4e363ab-5f3f-45a6-8840-03bc08acacba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.165101 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-dns-svc\") pod \"c4e363ab-5f3f-45a6-8840-03bc08acacba\" (UID: \"c4e363ab-5f3f-45a6-8840-03bc08acacba\") " Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.165498 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b309e015-8f4c-4c29-bf9c-7aa1e822c721-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.165581 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.165700 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4e363ab-5f3f-45a6-8840-03bc08acacba" (UID: "c4e363ab-5f3f-45a6-8840-03bc08acacba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.171013 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4e363ab-5f3f-45a6-8840-03bc08acacba-kube-api-access-g7x7x" (OuterVolumeSpecName: "kube-api-access-g7x7x") pod "c4e363ab-5f3f-45a6-8840-03bc08acacba" (UID: "c4e363ab-5f3f-45a6-8840-03bc08acacba"). InnerVolumeSpecName "kube-api-access-g7x7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.171267 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b309e015-8f4c-4c29-bf9c-7aa1e822c721-kube-api-access-vxf7t" (OuterVolumeSpecName: "kube-api-access-vxf7t") pod "b309e015-8f4c-4c29-bf9c-7aa1e822c721" (UID: "b309e015-8f4c-4c29-bf9c-7aa1e822c721"). InnerVolumeSpecName "kube-api-access-vxf7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.268421 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7x7x\" (UniqueName: \"kubernetes.io/projected/c4e363ab-5f3f-45a6-8840-03bc08acacba-kube-api-access-g7x7x\") on node \"crc\" DevicePath \"\"" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.268545 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4e363ab-5f3f-45a6-8840-03bc08acacba-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.268581 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxf7t\" (UniqueName: \"kubernetes.io/projected/b309e015-8f4c-4c29-bf9c-7aa1e822c721-kube-api-access-vxf7t\") on node \"crc\" DevicePath \"\"" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.493312 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c","Type":"ContainerStarted","Data":"9c0961e4c5c1b380e25b4b984eda2092f6a4523489dc2ab507510ae673855d7a"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.498286 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" event={"ID":"9e62dafe-2ab3-467f-adda-d18766ddd450","Type":"ContainerStarted","Data":"2790fb9ce8289471d956a6a24c47b65328923fed0f5f4a1c6f45e886c788dcf4"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.498331 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.502540 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n84gv" event={"ID":"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6","Type":"ContainerStarted","Data":"f92de9a43d650c8d08e15516d369a71361f2a1c79f1946739aef031801b9812b"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.505751 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dcf4b178-75ee-47fb-81e4-e6b277eb1790","Type":"ContainerStarted","Data":"0b44a0731057457a37fc02194611bb667c8fdefa9a770631b91b70d997c40b65"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.507399 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a6c91be1-1707-4e13-a2ba-4d9cf48d9622","Type":"ContainerStarted","Data":"85878efb48481921ef0aac03f18cf613ad02ee52d29dc378e8bbffe20ab634f6"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.508778 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c","Type":"ContainerStarted","Data":"e1fb1fc56de484f2a92aeafcc986d51404f6a3b8e03753d1b05a2d8adf62ce76"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.512399 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" event={"ID":"c4e363ab-5f3f-45a6-8840-03bc08acacba","Type":"ContainerDied","Data":"c8a4fddf10e37ec4cfc982eb9aa34a704cdc874dcad23e180c9350f5fd822e8c"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.512463 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vfvzm" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.522019 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c25a6d1-ee44-438c-a16d-e6952cfda420","Type":"ContainerStarted","Data":"64d5acfa5ac3affc79512c1007256b05a19ec5a3e209fe3a91578ee2901b87b3"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.522658 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" podStartSLOduration=2.683675388 podStartE2EDuration="16.522638369s" podCreationTimestamp="2025-10-10 15:03:39 +0000 UTC" firstStartedPulling="2025-10-10 15:03:40.026506207 +0000 UTC m=+1122.476221755" lastFinishedPulling="2025-10-10 15:03:53.865469188 +0000 UTC m=+1136.315184736" observedRunningTime="2025-10-10 15:03:55.520266105 +0000 UTC m=+1137.969981643" watchObservedRunningTime="2025-10-10 15:03:55.522638369 +0000 UTC m=+1137.972353917" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.525204 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e","Type":"ContainerStarted","Data":"9c05966ceccac74c34fc128399f37a6a0c27ea686de2c2efe9f4b3d85242332d"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.528855 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.530439 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-6lc8q" event={"ID":"b309e015-8f4c-4c29-bf9c-7aa1e822c721","Type":"ContainerDied","Data":"2b61c5bcd4a8e5adace3fe8819de2c21f1776afa2f8d9aa98f3feb6091fd7ea5"} Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.658238 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vfvzm"] Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.668244 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vfvzm"] Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.695715 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6lc8q"] Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.700335 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6lc8q"] Oct 10 15:03:55 crc kubenswrapper[4788]: I1010 15:03:55.704613 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 15:03:55 crc kubenswrapper[4788]: W1010 15:03:55.822121 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b927859_0238_46f5_90aa_4afa094c434e.slice/crio-4e9b28eae87eb831d3e9d85e3e537cebf371ca67643b93c002ab2145466e3187 WatchSource:0}: Error finding container 4e9b28eae87eb831d3e9d85e3e537cebf371ca67643b93c002ab2145466e3187: Status 404 returned error can't find the container with id 4e9b28eae87eb831d3e9d85e3e537cebf371ca67643b93c002ab2145466e3187 Oct 10 15:03:56 crc kubenswrapper[4788]: I1010 15:03:56.252678 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b309e015-8f4c-4c29-bf9c-7aa1e822c721" path="/var/lib/kubelet/pods/b309e015-8f4c-4c29-bf9c-7aa1e822c721/volumes" Oct 10 15:03:56 crc kubenswrapper[4788]: I1010 15:03:56.255971 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4e363ab-5f3f-45a6-8840-03bc08acacba" path="/var/lib/kubelet/pods/c4e363ab-5f3f-45a6-8840-03bc08acacba/volumes" Oct 10 15:03:56 crc kubenswrapper[4788]: I1010 15:03:56.547859 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b927859-0238-46f5-90aa-4afa094c434e","Type":"ContainerStarted","Data":"4e9b28eae87eb831d3e9d85e3e537cebf371ca67643b93c002ab2145466e3187"} Oct 10 15:03:56 crc kubenswrapper[4788]: I1010 15:03:56.746348 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-wpgvn"] Oct 10 15:03:57 crc kubenswrapper[4788]: W1010 15:03:57.284791 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6627a65b_765d_430c_99f5_eddd285d4289.slice/crio-e1ee3a66537f2628e8091a67b796dae5615ec3641c021bf5c3914a84f173fe37 WatchSource:0}: Error finding container e1ee3a66537f2628e8091a67b796dae5615ec3641c021bf5c3914a84f173fe37: Status 404 returned error can't find the container with id e1ee3a66537f2628e8091a67b796dae5615ec3641c021bf5c3914a84f173fe37 Oct 10 15:03:57 crc kubenswrapper[4788]: I1010 15:03:57.562751 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wpgvn" event={"ID":"6627a65b-765d-430c-99f5-eddd285d4289","Type":"ContainerStarted","Data":"e1ee3a66537f2628e8091a67b796dae5615ec3641c021bf5c3914a84f173fe37"} Oct 10 15:03:59 crc kubenswrapper[4788]: I1010 15:03:59.446427 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:03:59 crc kubenswrapper[4788]: I1010 15:03:59.501172 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jfxp2"] Oct 10 15:04:03 crc kubenswrapper[4788]: I1010 15:04:03.624738 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4075bf4d-01ed-4d7c-9fb8-576dd2866821","Type":"ContainerStarted","Data":"2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9"} Oct 10 15:04:03 crc kubenswrapper[4788]: I1010 15:04:03.625488 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 10 15:04:03 crc kubenswrapper[4788]: I1010 15:04:03.629791 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" event={"ID":"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a","Type":"ContainerStarted","Data":"ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075"} Oct 10 15:04:03 crc kubenswrapper[4788]: I1010 15:04:03.629928 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" podUID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" containerName="dnsmasq-dns" containerID="cri-o://ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075" gracePeriod=10 Oct 10 15:04:03 crc kubenswrapper[4788]: I1010 15:04:03.629939 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:04:03 crc kubenswrapper[4788]: I1010 15:04:03.647769 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.598186298 podStartE2EDuration="20.647747609s" podCreationTimestamp="2025-10-10 15:03:43 +0000 UTC" firstStartedPulling="2025-10-10 15:03:54.386687211 +0000 UTC m=+1136.836402769" lastFinishedPulling="2025-10-10 15:04:02.436248532 +0000 UTC m=+1144.885964080" observedRunningTime="2025-10-10 15:04:03.643372291 +0000 UTC m=+1146.093087839" watchObservedRunningTime="2025-10-10 15:04:03.647747609 +0000 UTC m=+1146.097463157" Oct 10 15:04:03 crc kubenswrapper[4788]: I1010 15:04:03.668812 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" podStartSLOduration=11.51371117 podStartE2EDuration="25.668745796s" podCreationTimestamp="2025-10-10 15:03:38 +0000 UTC" firstStartedPulling="2025-10-10 15:03:39.716520217 +0000 UTC m=+1122.166235765" lastFinishedPulling="2025-10-10 15:03:53.871554843 +0000 UTC m=+1136.321270391" observedRunningTime="2025-10-10 15:04:03.665604142 +0000 UTC m=+1146.115319690" watchObservedRunningTime="2025-10-10 15:04:03.668745796 +0000 UTC m=+1146.118461344" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.060133 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.144006 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-config\") pod \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.144185 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-dns-svc\") pod \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.144246 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r25nd\" (UniqueName: \"kubernetes.io/projected/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-kube-api-access-r25nd\") pod \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\" (UID: \"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a\") " Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.150610 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-kube-api-access-r25nd" (OuterVolumeSpecName: "kube-api-access-r25nd") pod "6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" (UID: "6bc95fb6-a85a-440d-8fcd-62386c6a3b5a"). InnerVolumeSpecName "kube-api-access-r25nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.216603 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-config" (OuterVolumeSpecName: "config") pod "6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" (UID: "6bc95fb6-a85a-440d-8fcd-62386c6a3b5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.216724 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" (UID: "6bc95fb6-a85a-440d-8fcd-62386c6a3b5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.245643 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.245675 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.245687 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r25nd\" (UniqueName: \"kubernetes.io/projected/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a-kube-api-access-r25nd\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.640971 4788 generic.go:334] "Generic (PLEG): container finished" podID="6627a65b-765d-430c-99f5-eddd285d4289" containerID="1c8f0a077f21d24aafc30fb99c1d2ad328ddb6a3998fad99ac9bedd96c095da7" exitCode=0 Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.641094 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wpgvn" event={"ID":"6627a65b-765d-430c-99f5-eddd285d4289","Type":"ContainerDied","Data":"1c8f0a077f21d24aafc30fb99c1d2ad328ddb6a3998fad99ac9bedd96c095da7"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.645225 4788 generic.go:334] "Generic (PLEG): container finished" podID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" containerID="ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075" exitCode=0 Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.645287 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" event={"ID":"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a","Type":"ContainerDied","Data":"ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.645324 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" event={"ID":"6bc95fb6-a85a-440d-8fcd-62386c6a3b5a","Type":"ContainerDied","Data":"3ebb16c38bd727fba2bab9c66dea87220408b0f4286ab515910f447d62cc8bef"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.645398 4788 scope.go:117] "RemoveContainer" containerID="ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.645553 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jfxp2" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.648612 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n84gv" event={"ID":"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6","Type":"ContainerStarted","Data":"94627d4d5603047c40017358c990871e5690c6c59e30cc3be499dae8b6bf6332"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.651656 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a6c91be1-1707-4e13-a2ba-4d9cf48d9622","Type":"ContainerStarted","Data":"ec7fa8d6b4673ceda9dd84c9e935b5a50e9b9b7c096828ed3a77bd95edb51e09"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.655835 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e","Type":"ContainerStarted","Data":"507ce22908f8278601f3c702001bc28ce6a57e91e11224453c2ba8d3538b36de"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.657648 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c","Type":"ContainerStarted","Data":"6dfca80283e228ab25b825ce8120c260fb7c568dd80ba2c02f1a941cc12700a6"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.658551 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.668899 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b927859-0238-46f5-90aa-4afa094c434e","Type":"ContainerStarted","Data":"5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.674084 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c25a6d1-ee44-438c-a16d-e6952cfda420","Type":"ContainerStarted","Data":"154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20"} Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.710951 4788 scope.go:117] "RemoveContainer" containerID="eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.716475 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jfxp2"] Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.731386 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jfxp2"] Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.737300 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.954008505000001 podStartE2EDuration="19.737260404s" podCreationTimestamp="2025-10-10 15:03:45 +0000 UTC" firstStartedPulling="2025-10-10 15:03:54.724516391 +0000 UTC m=+1137.174231939" lastFinishedPulling="2025-10-10 15:04:03.50776827 +0000 UTC m=+1145.957483838" observedRunningTime="2025-10-10 15:04:04.724284294 +0000 UTC m=+1147.173999842" watchObservedRunningTime="2025-10-10 15:04:04.737260404 +0000 UTC m=+1147.186975942" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.751757 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-n84gv" podStartSLOduration=8.506141822 podStartE2EDuration="16.751738155s" podCreationTimestamp="2025-10-10 15:03:48 +0000 UTC" firstStartedPulling="2025-10-10 15:03:54.739820085 +0000 UTC m=+1137.189535623" lastFinishedPulling="2025-10-10 15:04:02.985416418 +0000 UTC m=+1145.435131956" observedRunningTime="2025-10-10 15:04:04.744935122 +0000 UTC m=+1147.194650670" watchObservedRunningTime="2025-10-10 15:04:04.751738155 +0000 UTC m=+1147.201453703" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.812946 4788 scope.go:117] "RemoveContainer" containerID="ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075" Oct 10 15:04:04 crc kubenswrapper[4788]: E1010 15:04:04.813997 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075\": container with ID starting with ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075 not found: ID does not exist" containerID="ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.814039 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075"} err="failed to get container status \"ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075\": rpc error: code = NotFound desc = could not find container \"ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075\": container with ID starting with ce132aaccd69b60cd3a3e2f212253ba9dafcf6c02cc1f9ba722eb6d8c62dc075 not found: ID does not exist" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.814066 4788 scope.go:117] "RemoveContainer" containerID="eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc" Oct 10 15:04:04 crc kubenswrapper[4788]: E1010 15:04:04.814668 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc\": container with ID starting with eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc not found: ID does not exist" containerID="eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc" Oct 10 15:04:04 crc kubenswrapper[4788]: I1010 15:04:04.814803 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc"} err="failed to get container status \"eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc\": rpc error: code = NotFound desc = could not find container \"eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc\": container with ID starting with eeef1f7cd4410cb74193cab1fdcfd41c72665ff88f757fab3572371d8018a2fc not found: ID does not exist" Oct 10 15:04:05 crc kubenswrapper[4788]: I1010 15:04:05.682570 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wpgvn" event={"ID":"6627a65b-765d-430c-99f5-eddd285d4289","Type":"ContainerStarted","Data":"559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9"} Oct 10 15:04:05 crc kubenswrapper[4788]: I1010 15:04:05.683000 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wpgvn" event={"ID":"6627a65b-765d-430c-99f5-eddd285d4289","Type":"ContainerStarted","Data":"310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92"} Oct 10 15:04:05 crc kubenswrapper[4788]: I1010 15:04:05.683349 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:04:05 crc kubenswrapper[4788]: I1010 15:04:05.683386 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:04:05 crc kubenswrapper[4788]: I1010 15:04:05.688050 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dcf4b178-75ee-47fb-81e4-e6b277eb1790","Type":"ContainerStarted","Data":"a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290"} Oct 10 15:04:05 crc kubenswrapper[4788]: I1010 15:04:05.690286 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c","Type":"ContainerStarted","Data":"54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45"} Oct 10 15:04:05 crc kubenswrapper[4788]: I1010 15:04:05.691345 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-n84gv" Oct 10 15:04:05 crc kubenswrapper[4788]: I1010 15:04:05.707886 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-wpgvn" podStartSLOduration=12.109920916 podStartE2EDuration="17.707860488s" podCreationTimestamp="2025-10-10 15:03:48 +0000 UTC" firstStartedPulling="2025-10-10 15:03:57.287833596 +0000 UTC m=+1139.737549144" lastFinishedPulling="2025-10-10 15:04:02.885773128 +0000 UTC m=+1145.335488716" observedRunningTime="2025-10-10 15:04:05.705898245 +0000 UTC m=+1148.155613793" watchObservedRunningTime="2025-10-10 15:04:05.707860488 +0000 UTC m=+1148.157576036" Oct 10 15:04:06 crc kubenswrapper[4788]: I1010 15:04:06.247711 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" path="/var/lib/kubelet/pods/6bc95fb6-a85a-440d-8fcd-62386c6a3b5a/volumes" Oct 10 15:04:07 crc kubenswrapper[4788]: I1010 15:04:07.723358 4788 generic.go:334] "Generic (PLEG): container finished" podID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerID="154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20" exitCode=0 Oct 10 15:04:07 crc kubenswrapper[4788]: I1010 15:04:07.723478 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c25a6d1-ee44-438c-a16d-e6952cfda420","Type":"ContainerDied","Data":"154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20"} Oct 10 15:04:07 crc kubenswrapper[4788]: I1010 15:04:07.728478 4788 generic.go:334] "Generic (PLEG): container finished" podID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" containerID="ec7fa8d6b4673ceda9dd84c9e935b5a50e9b9b7c096828ed3a77bd95edb51e09" exitCode=0 Oct 10 15:04:07 crc kubenswrapper[4788]: I1010 15:04:07.728568 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a6c91be1-1707-4e13-a2ba-4d9cf48d9622","Type":"ContainerDied","Data":"ec7fa8d6b4673ceda9dd84c9e935b5a50e9b9b7c096828ed3a77bd95edb51e09"} Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.534327 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.742664 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c25a6d1-ee44-438c-a16d-e6952cfda420","Type":"ContainerStarted","Data":"36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74"} Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.745316 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a6c91be1-1707-4e13-a2ba-4d9cf48d9622","Type":"ContainerStarted","Data":"06b8e0b90057fe848ada9d1d4fb30735ed3b1c3365172a8e3399572d5e67a2a7"} Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.747529 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e","Type":"ContainerStarted","Data":"f363963ef14d25b3ce98764062c7aee163273743e8ae4d349c9bc2b1bd5df94f"} Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.749402 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b927859-0238-46f5-90aa-4afa094c434e","Type":"ContainerStarted","Data":"61dd4ebbf802647ebbc1de9086c92e02600921dcbef187e7ed7f85e677087139"} Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.772775 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.49679869 podStartE2EDuration="28.772739513s" podCreationTimestamp="2025-10-10 15:03:40 +0000 UTC" firstStartedPulling="2025-10-10 15:03:54.726533745 +0000 UTC m=+1137.176249293" lastFinishedPulling="2025-10-10 15:04:03.002474568 +0000 UTC m=+1145.452190116" observedRunningTime="2025-10-10 15:04:08.768316114 +0000 UTC m=+1151.218031662" watchObservedRunningTime="2025-10-10 15:04:08.772739513 +0000 UTC m=+1151.222455061" Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.796506 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.603179003 podStartE2EDuration="20.796476584s" podCreationTimestamp="2025-10-10 15:03:48 +0000 UTC" firstStartedPulling="2025-10-10 15:03:54.805769425 +0000 UTC m=+1137.255484973" lastFinishedPulling="2025-10-10 15:04:07.999067006 +0000 UTC m=+1150.448782554" observedRunningTime="2025-10-10 15:04:08.786112374 +0000 UTC m=+1151.235827922" watchObservedRunningTime="2025-10-10 15:04:08.796476584 +0000 UTC m=+1151.246192152" Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.820902 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=19.40159618 podStartE2EDuration="27.820878733s" podCreationTimestamp="2025-10-10 15:03:41 +0000 UTC" firstStartedPulling="2025-10-10 15:03:54.466174897 +0000 UTC m=+1136.915890435" lastFinishedPulling="2025-10-10 15:04:02.88545744 +0000 UTC m=+1145.335172988" observedRunningTime="2025-10-10 15:04:08.815699923 +0000 UTC m=+1151.265415481" watchObservedRunningTime="2025-10-10 15:04:08.820878733 +0000 UTC m=+1151.270594281" Oct 10 15:04:08 crc kubenswrapper[4788]: I1010 15:04:08.839043 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.706402659 podStartE2EDuration="17.839022843s" podCreationTimestamp="2025-10-10 15:03:51 +0000 UTC" firstStartedPulling="2025-10-10 15:03:55.825900577 +0000 UTC m=+1138.275616125" lastFinishedPulling="2025-10-10 15:04:07.958520751 +0000 UTC m=+1150.408236309" observedRunningTime="2025-10-10 15:04:08.831434488 +0000 UTC m=+1151.281150036" watchObservedRunningTime="2025-10-10 15:04:08.839022843 +0000 UTC m=+1151.288738391" Oct 10 15:04:09 crc kubenswrapper[4788]: I1010 15:04:09.565363 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 10 15:04:10 crc kubenswrapper[4788]: I1010 15:04:10.565107 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 10 15:04:10 crc kubenswrapper[4788]: I1010 15:04:10.615605 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 10 15:04:10 crc kubenswrapper[4788]: I1010 15:04:10.715800 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 10 15:04:10 crc kubenswrapper[4788]: I1010 15:04:10.772238 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 10 15:04:10 crc kubenswrapper[4788]: I1010 15:04:10.775252 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 10 15:04:10 crc kubenswrapper[4788]: I1010 15:04:10.821045 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 10 15:04:10 crc kubenswrapper[4788]: I1010 15:04:10.827242 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.173573 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f2rcn"] Oct 10 15:04:11 crc kubenswrapper[4788]: E1010 15:04:11.174024 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" containerName="init" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.174042 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" containerName="init" Oct 10 15:04:11 crc kubenswrapper[4788]: E1010 15:04:11.174063 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" containerName="dnsmasq-dns" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.174069 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" containerName="dnsmasq-dns" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.174337 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bc95fb6-a85a-440d-8fcd-62386c6a3b5a" containerName="dnsmasq-dns" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.175443 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.178846 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.186599 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f2rcn"] Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.245316 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7m7sz"] Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.246854 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.249331 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.264263 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7m7sz"] Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.275428 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.275522 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8jl9\" (UniqueName: \"kubernetes.io/projected/0fd8ad49-e401-4efb-839a-947769f4d9f9-kube-api-access-w8jl9\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.275575 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-config\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.275601 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.327160 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f2rcn"] Oct 10 15:04:11 crc kubenswrapper[4788]: E1010 15:04:11.330179 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-w8jl9 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" podUID="0fd8ad49-e401-4efb-839a-947769f4d9f9" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.377621 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.377697 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-combined-ca-bundle\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.377732 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kph2n"] Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.377765 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8jl9\" (UniqueName: \"kubernetes.io/projected/0fd8ad49-e401-4efb-839a-947769f4d9f9-kube-api-access-w8jl9\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.378479 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-config\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.378535 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.378651 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovn-rundir\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.378761 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s656\" (UniqueName: \"kubernetes.io/projected/953774da-73b1-4f46-aa87-8deba0bcc077-kube-api-access-5s656\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.378813 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953774da-73b1-4f46-aa87-8deba0bcc077-config\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.378960 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovs-rundir\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.378986 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.379200 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.379628 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-config\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.380804 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.380977 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.387444 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.396275 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kph2n"] Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.402686 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.405935 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.414587 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.415053 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-9gh2r" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.415320 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.416565 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.463680 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481074 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481126 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njxxh\" (UniqueName: \"kubernetes.io/projected/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-kube-api-access-njxxh\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481172 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55n9j\" (UniqueName: \"kubernetes.io/projected/a9adb465-b660-4d2c-bae8-f0b8024761b8-kube-api-access-55n9j\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481196 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481225 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s656\" (UniqueName: \"kubernetes.io/projected/953774da-73b1-4f46-aa87-8deba0bcc077-kube-api-access-5s656\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481248 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481281 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953774da-73b1-4f46-aa87-8deba0bcc077-config\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481308 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-config\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481344 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovs-rundir\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481384 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481419 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-combined-ca-bundle\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481450 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481477 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-scripts\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481511 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481547 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481589 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481613 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-config\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.481635 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovn-rundir\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.482027 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovn-rundir\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.482666 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovs-rundir\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.484593 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953774da-73b1-4f46-aa87-8deba0bcc077-config\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.487364 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-combined-ca-bundle\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.497329 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8jl9\" (UniqueName: \"kubernetes.io/projected/0fd8ad49-e401-4efb-839a-947769f4d9f9-kube-api-access-w8jl9\") pod \"dnsmasq-dns-7fd796d7df-f2rcn\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.499064 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s656\" (UniqueName: \"kubernetes.io/projected/953774da-73b1-4f46-aa87-8deba0bcc077-kube-api-access-5s656\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.520693 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7m7sz\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.568153 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.583743 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.583820 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-scripts\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.583857 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.583903 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.583972 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.584002 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-config\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.584041 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.584066 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njxxh\" (UniqueName: \"kubernetes.io/projected/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-kube-api-access-njxxh\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.584085 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55n9j\" (UniqueName: \"kubernetes.io/projected/a9adb465-b660-4d2c-bae8-f0b8024761b8-kube-api-access-55n9j\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.584108 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.584129 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.584188 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-config\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.584720 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.585118 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.586179 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-config\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.586316 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.586404 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-config\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.586444 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.586982 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-scripts\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.589839 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.592090 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.599234 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.606116 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njxxh\" (UniqueName: \"kubernetes.io/projected/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-kube-api-access-njxxh\") pod \"dnsmasq-dns-86db49b7ff-kph2n\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.607697 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55n9j\" (UniqueName: \"kubernetes.io/projected/a9adb465-b660-4d2c-bae8-f0b8024761b8-kube-api-access-55n9j\") pod \"ovn-northd-0\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.703348 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.773965 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.790954 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.823268 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.865671 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.867268 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.889876 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-dns-svc\") pod \"0fd8ad49-e401-4efb-839a-947769f4d9f9\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.889934 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-ovsdbserver-nb\") pod \"0fd8ad49-e401-4efb-839a-947769f4d9f9\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.889986 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-config\") pod \"0fd8ad49-e401-4efb-839a-947769f4d9f9\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.890061 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8jl9\" (UniqueName: \"kubernetes.io/projected/0fd8ad49-e401-4efb-839a-947769f4d9f9-kube-api-access-w8jl9\") pod \"0fd8ad49-e401-4efb-839a-947769f4d9f9\" (UID: \"0fd8ad49-e401-4efb-839a-947769f4d9f9\") " Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.890839 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0fd8ad49-e401-4efb-839a-947769f4d9f9" (UID: "0fd8ad49-e401-4efb-839a-947769f4d9f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.890868 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-config" (OuterVolumeSpecName: "config") pod "0fd8ad49-e401-4efb-839a-947769f4d9f9" (UID: "0fd8ad49-e401-4efb-839a-947769f4d9f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.892629 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0fd8ad49-e401-4efb-839a-947769f4d9f9" (UID: "0fd8ad49-e401-4efb-839a-947769f4d9f9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.902460 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fd8ad49-e401-4efb-839a-947769f4d9f9-kube-api-access-w8jl9" (OuterVolumeSpecName: "kube-api-access-w8jl9") pod "0fd8ad49-e401-4efb-839a-947769f4d9f9" (UID: "0fd8ad49-e401-4efb-839a-947769f4d9f9"). InnerVolumeSpecName "kube-api-access-w8jl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.992427 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.992500 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.992519 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fd8ad49-e401-4efb-839a-947769f4d9f9-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:11 crc kubenswrapper[4788]: I1010 15:04:11.992535 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8jl9\" (UniqueName: \"kubernetes.io/projected/0fd8ad49-e401-4efb-839a-947769f4d9f9-kube-api-access-w8jl9\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.035382 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7m7sz"] Oct 10 15:04:12 crc kubenswrapper[4788]: W1010 15:04:12.044456 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod953774da_73b1_4f46_aa87_8deba0bcc077.slice/crio-6fe04936c82a41cfa98096d1c0aafacbbe3c2c071e21ef8ff5324af257d2fe60 WatchSource:0}: Error finding container 6fe04936c82a41cfa98096d1c0aafacbbe3c2c071e21ef8ff5324af257d2fe60: Status 404 returned error can't find the container with id 6fe04936c82a41cfa98096d1c0aafacbbe3c2c071e21ef8ff5324af257d2fe60 Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.181489 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kph2n"] Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.282610 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 15:04:12 crc kubenswrapper[4788]: W1010 15:04:12.287497 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9adb465_b660_4d2c_bae8_f0b8024761b8.slice/crio-e2e5e941f3144ddea505b482e85894d2ab1422c68a655bf2103d1f36b70fffd7 WatchSource:0}: Error finding container e2e5e941f3144ddea505b482e85894d2ab1422c68a655bf2103d1f36b70fffd7: Status 404 returned error can't find the container with id e2e5e941f3144ddea505b482e85894d2ab1422c68a655bf2103d1f36b70fffd7 Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.801481 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" event={"ID":"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea","Type":"ContainerStarted","Data":"a12e79a85279a866d0d032554e97973ad8279573c19773bd16b3e7cc17f0e4be"} Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.802668 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7m7sz" event={"ID":"953774da-73b1-4f46-aa87-8deba0bcc077","Type":"ContainerStarted","Data":"6fe04936c82a41cfa98096d1c0aafacbbe3c2c071e21ef8ff5324af257d2fe60"} Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.803928 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a9adb465-b660-4d2c-bae8-f0b8024761b8","Type":"ContainerStarted","Data":"e2e5e941f3144ddea505b482e85894d2ab1422c68a655bf2103d1f36b70fffd7"} Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.803963 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-f2rcn" Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.851426 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f2rcn"] Oct 10 15:04:12 crc kubenswrapper[4788]: I1010 15:04:12.857090 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-f2rcn"] Oct 10 15:04:13 crc kubenswrapper[4788]: I1010 15:04:13.142658 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 10 15:04:13 crc kubenswrapper[4788]: I1010 15:04:13.142723 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 10 15:04:14 crc kubenswrapper[4788]: I1010 15:04:14.251089 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fd8ad49-e401-4efb-839a-947769f4d9f9" path="/var/lib/kubelet/pods/0fd8ad49-e401-4efb-839a-947769f4d9f9/volumes" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.416439 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.556287 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kph2n"] Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.599420 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-zjjvp"] Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.600998 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.620820 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zjjvp"] Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.658937 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crkwp\" (UniqueName: \"kubernetes.io/projected/3f57f828-9ffd-4caf-872a-fc3fd1913df5-kube-api-access-crkwp\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.659005 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.659192 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.659318 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-config\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.659385 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-dns-svc\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.760969 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-config\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.761052 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-dns-svc\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.761133 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crkwp\" (UniqueName: \"kubernetes.io/projected/3f57f828-9ffd-4caf-872a-fc3fd1913df5-kube-api-access-crkwp\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.761249 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.761294 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.762158 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-config\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.762461 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-dns-svc\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.762499 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.762737 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.789118 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crkwp\" (UniqueName: \"kubernetes.io/projected/3f57f828-9ffd-4caf-872a-fc3fd1913df5-kube-api-access-crkwp\") pod \"dnsmasq-dns-698758b865-zjjvp\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.832247 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7m7sz" event={"ID":"953774da-73b1-4f46-aa87-8deba0bcc077","Type":"ContainerStarted","Data":"69a85b38a0e4154979d9d710277ca0a1c7a9d6c594026db5c938404d7dc3173b"} Oct 10 15:04:15 crc kubenswrapper[4788]: I1010 15:04:15.919004 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.307267 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zjjvp"] Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.750990 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.767336 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.768487 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.771096 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.771199 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-sjvwb" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.772286 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.773697 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.840954 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zjjvp" event={"ID":"3f57f828-9ffd-4caf-872a-fc3fd1913df5","Type":"ContainerStarted","Data":"bf083e2f2ea224805a96fe73d8de4e530cad6e3ec5a2d1698ebb229b1d4eaecc"} Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.890245 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-lock\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.890349 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.890371 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.890423 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-cache\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.890494 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vl8d\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-kube-api-access-6vl8d\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.992125 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vl8d\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-kube-api-access-6vl8d\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.992214 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-lock\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.992272 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.992297 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.992348 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-cache\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: E1010 15:04:16.992525 4788 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 15:04:16 crc kubenswrapper[4788]: E1010 15:04:16.992547 4788 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 15:04:16 crc kubenswrapper[4788]: E1010 15:04:16.992605 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift podName:edada8e6-c183-4ef8-bcd5-f5c617e7615e nodeName:}" failed. No retries permitted until 2025-10-10 15:04:17.492584162 +0000 UTC m=+1159.942299710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift") pod "swift-storage-0" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e") : configmap "swift-ring-files" not found Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.992696 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.992994 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-cache\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:16 crc kubenswrapper[4788]: I1010 15:04:16.993014 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-lock\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.034933 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vl8d\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-kube-api-access-6vl8d\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.041359 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.105024 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-h7nxx"] Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.106984 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.112635 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.113188 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.115504 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.119421 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-h7nxx"] Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.162790 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-h7nxx"] Oct 10 15:04:17 crc kubenswrapper[4788]: E1010 15:04:17.163733 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-r6l6v ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-r6l6v ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-h7nxx" podUID="73ebd7f6-2edb-4bdb-a926-0377c12dbc55" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.171951 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-fr4km"] Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.173391 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.207851 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-fr4km"] Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.297825 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-combined-ca-bundle\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.297883 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-etc-swift\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.297917 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-scripts\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.298196 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-ring-data-devices\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.298332 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-ring-data-devices\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.298580 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-swiftconf\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.298693 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6l6v\" (UniqueName: \"kubernetes.io/projected/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-kube-api-access-r6l6v\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.298778 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-scripts\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.298854 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-dispersionconf\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.298993 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-dispersionconf\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.299054 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-swiftconf\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.299103 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4h6c\" (UniqueName: \"kubernetes.io/projected/bb647ff7-3bd1-4325-a6a6-916a6f134fea-kube-api-access-x4h6c\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.299397 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-combined-ca-bundle\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.299545 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb647ff7-3bd1-4325-a6a6-916a6f134fea-etc-swift\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402008 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-ring-data-devices\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402084 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-ring-data-devices\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402168 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-swiftconf\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402213 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6l6v\" (UniqueName: \"kubernetes.io/projected/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-kube-api-access-r6l6v\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402265 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-scripts\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402292 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-dispersionconf\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402325 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-dispersionconf\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402350 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-swiftconf\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402375 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4h6c\" (UniqueName: \"kubernetes.io/projected/bb647ff7-3bd1-4325-a6a6-916a6f134fea-kube-api-access-x4h6c\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402440 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-combined-ca-bundle\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402482 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb647ff7-3bd1-4325-a6a6-916a6f134fea-etc-swift\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402521 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-combined-ca-bundle\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402551 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-etc-swift\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.402592 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-scripts\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.403674 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-scripts\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.404377 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-ring-data-devices\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.404954 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-ring-data-devices\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.406188 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb647ff7-3bd1-4325-a6a6-916a6f134fea-etc-swift\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.406312 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-etc-swift\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.406454 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-scripts\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.410897 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-dispersionconf\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.410897 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-swiftconf\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.411301 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-swiftconf\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.411717 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-dispersionconf\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.421759 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-combined-ca-bundle\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.423973 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-combined-ca-bundle\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.425409 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6l6v\" (UniqueName: \"kubernetes.io/projected/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-kube-api-access-r6l6v\") pod \"swift-ring-rebalance-h7nxx\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.432359 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4h6c\" (UniqueName: \"kubernetes.io/projected/bb647ff7-3bd1-4325-a6a6-916a6f134fea-kube-api-access-x4h6c\") pod \"swift-ring-rebalance-fr4km\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.497951 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.506831 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:17 crc kubenswrapper[4788]: E1010 15:04:17.507319 4788 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 15:04:17 crc kubenswrapper[4788]: E1010 15:04:17.507403 4788 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 15:04:17 crc kubenswrapper[4788]: E1010 15:04:17.507543 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift podName:edada8e6-c183-4ef8-bcd5-f5c617e7615e nodeName:}" failed. No retries permitted until 2025-10-10 15:04:18.507497713 +0000 UTC m=+1160.957213301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift") pod "swift-storage-0" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e") : configmap "swift-ring-files" not found Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.776044 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-fr4km"] Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.853499 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.854059 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fr4km" event={"ID":"bb647ff7-3bd1-4325-a6a6-916a6f134fea","Type":"ContainerStarted","Data":"833b0c2dded97b1a0bcf5b98261e897597ff700f911b865b6a640d1223a007dc"} Oct 10 15:04:17 crc kubenswrapper[4788]: I1010 15:04:17.865796 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.018553 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-combined-ca-bundle\") pod \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.018661 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-dispersionconf\") pod \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.018686 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-scripts\") pod \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.018742 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6l6v\" (UniqueName: \"kubernetes.io/projected/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-kube-api-access-r6l6v\") pod \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.018772 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-etc-swift\") pod \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.018823 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-swiftconf\") pod \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.018840 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-ring-data-devices\") pod \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\" (UID: \"73ebd7f6-2edb-4bdb-a926-0377c12dbc55\") " Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.019519 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "73ebd7f6-2edb-4bdb-a926-0377c12dbc55" (UID: "73ebd7f6-2edb-4bdb-a926-0377c12dbc55"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.019762 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-scripts" (OuterVolumeSpecName: "scripts") pod "73ebd7f6-2edb-4bdb-a926-0377c12dbc55" (UID: "73ebd7f6-2edb-4bdb-a926-0377c12dbc55"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.019771 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "73ebd7f6-2edb-4bdb-a926-0377c12dbc55" (UID: "73ebd7f6-2edb-4bdb-a926-0377c12dbc55"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.025617 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-kube-api-access-r6l6v" (OuterVolumeSpecName: "kube-api-access-r6l6v") pod "73ebd7f6-2edb-4bdb-a926-0377c12dbc55" (UID: "73ebd7f6-2edb-4bdb-a926-0377c12dbc55"). InnerVolumeSpecName "kube-api-access-r6l6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.025864 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "73ebd7f6-2edb-4bdb-a926-0377c12dbc55" (UID: "73ebd7f6-2edb-4bdb-a926-0377c12dbc55"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.027799 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "73ebd7f6-2edb-4bdb-a926-0377c12dbc55" (UID: "73ebd7f6-2edb-4bdb-a926-0377c12dbc55"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.031359 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73ebd7f6-2edb-4bdb-a926-0377c12dbc55" (UID: "73ebd7f6-2edb-4bdb-a926-0377c12dbc55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.120782 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6l6v\" (UniqueName: \"kubernetes.io/projected/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-kube-api-access-r6l6v\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.121123 4788 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.121356 4788 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.121422 4788 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.121482 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.121592 4788 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.121666 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ebd7f6-2edb-4bdb-a926-0377c12dbc55-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.528731 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:18 crc kubenswrapper[4788]: E1010 15:04:18.528902 4788 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 15:04:18 crc kubenswrapper[4788]: E1010 15:04:18.529294 4788 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 15:04:18 crc kubenswrapper[4788]: E1010 15:04:18.529460 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift podName:edada8e6-c183-4ef8-bcd5-f5c617e7615e nodeName:}" failed. No retries permitted until 2025-10-10 15:04:20.529336471 +0000 UTC m=+1162.979052029 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift") pod "swift-storage-0" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e") : configmap "swift-ring-files" not found Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.874191 4788 generic.go:334] "Generic (PLEG): container finished" podID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerID="b66badae53fe5050595a9fc745c9c05d52a914ddb1db479fee6dbe97f1709580" exitCode=0 Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.874585 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zjjvp" event={"ID":"3f57f828-9ffd-4caf-872a-fc3fd1913df5","Type":"ContainerDied","Data":"b66badae53fe5050595a9fc745c9c05d52a914ddb1db479fee6dbe97f1709580"} Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.883768 4788 generic.go:334] "Generic (PLEG): container finished" podID="73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" containerID="de1a79a325ae0082413f44b631787caa3909fb3999c43b74c4122e1b4927b3d9" exitCode=0 Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.883986 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" event={"ID":"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea","Type":"ContainerDied","Data":"de1a79a325ae0082413f44b631787caa3909fb3999c43b74c4122e1b4927b3d9"} Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.884638 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-h7nxx" Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.950192 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-h7nxx"] Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.957500 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-h7nxx"] Oct 10 15:04:18 crc kubenswrapper[4788]: E1010 15:04:18.966486 4788 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.129:42964->38.102.83.129:41727: write tcp 38.102.83.129:42964->38.102.83.129:41727: write: broken pipe Oct 10 15:04:18 crc kubenswrapper[4788]: I1010 15:04:18.991722 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7m7sz" podStartSLOduration=7.991702754 podStartE2EDuration="7.991702754s" podCreationTimestamp="2025-10-10 15:04:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:04:18.98342316 +0000 UTC m=+1161.433138708" watchObservedRunningTime="2025-10-10 15:04:18.991702754 +0000 UTC m=+1161.441418292" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.524179 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.655440 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-sb\") pod \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.655622 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-config\") pod \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.655777 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-dns-svc\") pod \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.655834 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njxxh\" (UniqueName: \"kubernetes.io/projected/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-kube-api-access-njxxh\") pod \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.656035 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-nb\") pod \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\" (UID: \"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea\") " Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.662536 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-kube-api-access-njxxh" (OuterVolumeSpecName: "kube-api-access-njxxh") pod "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" (UID: "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea"). InnerVolumeSpecName "kube-api-access-njxxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.682770 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-config" (OuterVolumeSpecName: "config") pod "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" (UID: "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.685558 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" (UID: "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.687218 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" (UID: "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.691411 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" (UID: "73fa6d0f-90de-4ad8-90cc-cf0db1f9abea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.761415 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.761456 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.761470 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njxxh\" (UniqueName: \"kubernetes.io/projected/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-kube-api-access-njxxh\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.761483 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.761495 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.895378 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zjjvp" event={"ID":"3f57f828-9ffd-4caf-872a-fc3fd1913df5","Type":"ContainerStarted","Data":"4ae05de9ee9dfbfd7f28da6127b6ad354efa3958ccfb741c52c9a125d68d2d1d"} Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.895618 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.900412 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a9adb465-b660-4d2c-bae8-f0b8024761b8","Type":"ContainerStarted","Data":"335000a0690e32f11a92cfd8dfc8fe055195d80c307ea823dc4da83c8237d39d"} Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.900441 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a9adb465-b660-4d2c-bae8-f0b8024761b8","Type":"ContainerStarted","Data":"ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754"} Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.900552 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.902209 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" event={"ID":"73fa6d0f-90de-4ad8-90cc-cf0db1f9abea","Type":"ContainerDied","Data":"a12e79a85279a866d0d032554e97973ad8279573c19773bd16b3e7cc17f0e4be"} Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.902277 4788 scope.go:117] "RemoveContainer" containerID="de1a79a325ae0082413f44b631787caa3909fb3999c43b74c4122e1b4927b3d9" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.902450 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kph2n" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.924551 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-zjjvp" podStartSLOduration=4.924531588 podStartE2EDuration="4.924531588s" podCreationTimestamp="2025-10-10 15:04:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:04:19.918354171 +0000 UTC m=+1162.368069729" watchObservedRunningTime="2025-10-10 15:04:19.924531588 +0000 UTC m=+1162.374247136" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.942462 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.797227807 podStartE2EDuration="8.942441482s" podCreationTimestamp="2025-10-10 15:04:11 +0000 UTC" firstStartedPulling="2025-10-10 15:04:12.293219509 +0000 UTC m=+1154.742935057" lastFinishedPulling="2025-10-10 15:04:19.438433184 +0000 UTC m=+1161.888148732" observedRunningTime="2025-10-10 15:04:19.940642273 +0000 UTC m=+1162.390357821" watchObservedRunningTime="2025-10-10 15:04:19.942441482 +0000 UTC m=+1162.392157030" Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.991440 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kph2n"] Oct 10 15:04:19 crc kubenswrapper[4788]: I1010 15:04:19.997690 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kph2n"] Oct 10 15:04:20 crc kubenswrapper[4788]: I1010 15:04:20.245741 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73ebd7f6-2edb-4bdb-a926-0377c12dbc55" path="/var/lib/kubelet/pods/73ebd7f6-2edb-4bdb-a926-0377c12dbc55/volumes" Oct 10 15:04:20 crc kubenswrapper[4788]: I1010 15:04:20.246559 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" path="/var/lib/kubelet/pods/73fa6d0f-90de-4ad8-90cc-cf0db1f9abea/volumes" Oct 10 15:04:20 crc kubenswrapper[4788]: I1010 15:04:20.344100 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 10 15:04:20 crc kubenswrapper[4788]: I1010 15:04:20.416022 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 10 15:04:20 crc kubenswrapper[4788]: I1010 15:04:20.588884 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:20 crc kubenswrapper[4788]: E1010 15:04:20.589173 4788 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 15:04:20 crc kubenswrapper[4788]: E1010 15:04:20.589204 4788 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 15:04:20 crc kubenswrapper[4788]: E1010 15:04:20.589305 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift podName:edada8e6-c183-4ef8-bcd5-f5c617e7615e nodeName:}" failed. No retries permitted until 2025-10-10 15:04:24.589271565 +0000 UTC m=+1167.038987133 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift") pod "swift-storage-0" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e") : configmap "swift-ring-files" not found Oct 10 15:04:22 crc kubenswrapper[4788]: I1010 15:04:22.021334 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 10 15:04:22 crc kubenswrapper[4788]: I1010 15:04:22.075788 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 10 15:04:22 crc kubenswrapper[4788]: I1010 15:04:22.957069 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fr4km" event={"ID":"bb647ff7-3bd1-4325-a6a6-916a6f134fea","Type":"ContainerStarted","Data":"3af2beb51561d3ef0612a1db993a775db975046929eb4d1f3edfa355eb9cecc3"} Oct 10 15:04:22 crc kubenswrapper[4788]: I1010 15:04:22.985758 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-fr4km" podStartSLOduration=1.326709981 podStartE2EDuration="5.985739534s" podCreationTimestamp="2025-10-10 15:04:17 +0000 UTC" firstStartedPulling="2025-10-10 15:04:17.787454042 +0000 UTC m=+1160.237169590" lastFinishedPulling="2025-10-10 15:04:22.446483585 +0000 UTC m=+1164.896199143" observedRunningTime="2025-10-10 15:04:22.981726526 +0000 UTC m=+1165.431442074" watchObservedRunningTime="2025-10-10 15:04:22.985739534 +0000 UTC m=+1165.435455082" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.320521 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-mml9f"] Oct 10 15:04:23 crc kubenswrapper[4788]: E1010 15:04:23.321073 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" containerName="init" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.321090 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" containerName="init" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.321317 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="73fa6d0f-90de-4ad8-90cc-cf0db1f9abea" containerName="init" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.322096 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mml9f" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.333744 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mml9f"] Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.407519 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-jmlx2"] Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.409361 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jmlx2" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.417272 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jmlx2"] Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.446248 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mvkq\" (UniqueName: \"kubernetes.io/projected/f12ed2f7-2d5e-4a57-bd17-59537de8065a-kube-api-access-5mvkq\") pod \"keystone-db-create-mml9f\" (UID: \"f12ed2f7-2d5e-4a57-bd17-59537de8065a\") " pod="openstack/keystone-db-create-mml9f" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.548398 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxw2x\" (UniqueName: \"kubernetes.io/projected/b9f621aa-8944-4699-b706-53c84c2e9fe5-kube-api-access-pxw2x\") pod \"placement-db-create-jmlx2\" (UID: \"b9f621aa-8944-4699-b706-53c84c2e9fe5\") " pod="openstack/placement-db-create-jmlx2" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.548577 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mvkq\" (UniqueName: \"kubernetes.io/projected/f12ed2f7-2d5e-4a57-bd17-59537de8065a-kube-api-access-5mvkq\") pod \"keystone-db-create-mml9f\" (UID: \"f12ed2f7-2d5e-4a57-bd17-59537de8065a\") " pod="openstack/keystone-db-create-mml9f" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.595570 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mvkq\" (UniqueName: \"kubernetes.io/projected/f12ed2f7-2d5e-4a57-bd17-59537de8065a-kube-api-access-5mvkq\") pod \"keystone-db-create-mml9f\" (UID: \"f12ed2f7-2d5e-4a57-bd17-59537de8065a\") " pod="openstack/keystone-db-create-mml9f" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.653867 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mml9f" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.654751 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxw2x\" (UniqueName: \"kubernetes.io/projected/b9f621aa-8944-4699-b706-53c84c2e9fe5-kube-api-access-pxw2x\") pod \"placement-db-create-jmlx2\" (UID: \"b9f621aa-8944-4699-b706-53c84c2e9fe5\") " pod="openstack/placement-db-create-jmlx2" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.690020 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxw2x\" (UniqueName: \"kubernetes.io/projected/b9f621aa-8944-4699-b706-53c84c2e9fe5-kube-api-access-pxw2x\") pod \"placement-db-create-jmlx2\" (UID: \"b9f621aa-8944-4699-b706-53c84c2e9fe5\") " pod="openstack/placement-db-create-jmlx2" Oct 10 15:04:23 crc kubenswrapper[4788]: I1010 15:04:23.737900 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jmlx2" Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.170591 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mml9f"] Oct 10 15:04:24 crc kubenswrapper[4788]: W1010 15:04:24.180524 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf12ed2f7_2d5e_4a57_bd17_59537de8065a.slice/crio-e1d5753a1ca2b95a6f3d3039cdbd2376b9e518ff71bd0678212ebd8bd6843e42 WatchSource:0}: Error finding container e1d5753a1ca2b95a6f3d3039cdbd2376b9e518ff71bd0678212ebd8bd6843e42: Status 404 returned error can't find the container with id e1d5753a1ca2b95a6f3d3039cdbd2376b9e518ff71bd0678212ebd8bd6843e42 Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.260780 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-jmlx2"] Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.678009 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:24 crc kubenswrapper[4788]: E1010 15:04:24.678541 4788 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 10 15:04:24 crc kubenswrapper[4788]: E1010 15:04:24.678633 4788 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 10 15:04:24 crc kubenswrapper[4788]: E1010 15:04:24.678820 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift podName:edada8e6-c183-4ef8-bcd5-f5c617e7615e nodeName:}" failed. No retries permitted until 2025-10-10 15:04:32.678780933 +0000 UTC m=+1175.128496511 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift") pod "swift-storage-0" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e") : configmap "swift-ring-files" not found Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.974760 4788 generic.go:334] "Generic (PLEG): container finished" podID="f12ed2f7-2d5e-4a57-bd17-59537de8065a" containerID="c74f38ffd4f07c84655a1a8102dedf197116202fc557e371bc029b11a0688100" exitCode=0 Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.974923 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mml9f" event={"ID":"f12ed2f7-2d5e-4a57-bd17-59537de8065a","Type":"ContainerDied","Data":"c74f38ffd4f07c84655a1a8102dedf197116202fc557e371bc029b11a0688100"} Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.975206 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mml9f" event={"ID":"f12ed2f7-2d5e-4a57-bd17-59537de8065a","Type":"ContainerStarted","Data":"e1d5753a1ca2b95a6f3d3039cdbd2376b9e518ff71bd0678212ebd8bd6843e42"} Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.979154 4788 generic.go:334] "Generic (PLEG): container finished" podID="b9f621aa-8944-4699-b706-53c84c2e9fe5" containerID="93cdd8df755829312c77164115af990c8b6b186683f6a2fa4048af1e3e662e81" exitCode=0 Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.979241 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jmlx2" event={"ID":"b9f621aa-8944-4699-b706-53c84c2e9fe5","Type":"ContainerDied","Data":"93cdd8df755829312c77164115af990c8b6b186683f6a2fa4048af1e3e662e81"} Oct 10 15:04:24 crc kubenswrapper[4788]: I1010 15:04:24.979554 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jmlx2" event={"ID":"b9f621aa-8944-4699-b706-53c84c2e9fe5","Type":"ContainerStarted","Data":"c63d78a1670da524e215d3cb8687ee060742ded10c37a8c8ec0095beeff0c94b"} Oct 10 15:04:25 crc kubenswrapper[4788]: I1010 15:04:25.921359 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.012313 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-r4mvh"] Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.012662 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" podUID="9e62dafe-2ab3-467f-adda-d18766ddd450" containerName="dnsmasq-dns" containerID="cri-o://2790fb9ce8289471d956a6a24c47b65328923fed0f5f4a1c6f45e886c788dcf4" gracePeriod=10 Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.477842 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jmlx2" Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.623084 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxw2x\" (UniqueName: \"kubernetes.io/projected/b9f621aa-8944-4699-b706-53c84c2e9fe5-kube-api-access-pxw2x\") pod \"b9f621aa-8944-4699-b706-53c84c2e9fe5\" (UID: \"b9f621aa-8944-4699-b706-53c84c2e9fe5\") " Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.631322 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9f621aa-8944-4699-b706-53c84c2e9fe5-kube-api-access-pxw2x" (OuterVolumeSpecName: "kube-api-access-pxw2x") pod "b9f621aa-8944-4699-b706-53c84c2e9fe5" (UID: "b9f621aa-8944-4699-b706-53c84c2e9fe5"). InnerVolumeSpecName "kube-api-access-pxw2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.634728 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mml9f" Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.724876 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mvkq\" (UniqueName: \"kubernetes.io/projected/f12ed2f7-2d5e-4a57-bd17-59537de8065a-kube-api-access-5mvkq\") pod \"f12ed2f7-2d5e-4a57-bd17-59537de8065a\" (UID: \"f12ed2f7-2d5e-4a57-bd17-59537de8065a\") " Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.725369 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxw2x\" (UniqueName: \"kubernetes.io/projected/b9f621aa-8944-4699-b706-53c84c2e9fe5-kube-api-access-pxw2x\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.728279 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f12ed2f7-2d5e-4a57-bd17-59537de8065a-kube-api-access-5mvkq" (OuterVolumeSpecName: "kube-api-access-5mvkq") pod "f12ed2f7-2d5e-4a57-bd17-59537de8065a" (UID: "f12ed2f7-2d5e-4a57-bd17-59537de8065a"). InnerVolumeSpecName "kube-api-access-5mvkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:26 crc kubenswrapper[4788]: I1010 15:04:26.826718 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mvkq\" (UniqueName: \"kubernetes.io/projected/f12ed2f7-2d5e-4a57-bd17-59537de8065a-kube-api-access-5mvkq\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.017810 4788 generic.go:334] "Generic (PLEG): container finished" podID="9e62dafe-2ab3-467f-adda-d18766ddd450" containerID="2790fb9ce8289471d956a6a24c47b65328923fed0f5f4a1c6f45e886c788dcf4" exitCode=0 Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.017882 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" event={"ID":"9e62dafe-2ab3-467f-adda-d18766ddd450","Type":"ContainerDied","Data":"2790fb9ce8289471d956a6a24c47b65328923fed0f5f4a1c6f45e886c788dcf4"} Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.020250 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-jmlx2" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.020246 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-jmlx2" event={"ID":"b9f621aa-8944-4699-b706-53c84c2e9fe5","Type":"ContainerDied","Data":"c63d78a1670da524e215d3cb8687ee060742ded10c37a8c8ec0095beeff0c94b"} Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.020363 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c63d78a1670da524e215d3cb8687ee060742ded10c37a8c8ec0095beeff0c94b" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.023022 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mml9f" event={"ID":"f12ed2f7-2d5e-4a57-bd17-59537de8065a","Type":"ContainerDied","Data":"e1d5753a1ca2b95a6f3d3039cdbd2376b9e518ff71bd0678212ebd8bd6843e42"} Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.023042 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1d5753a1ca2b95a6f3d3039cdbd2376b9e518ff71bd0678212ebd8bd6843e42" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.023092 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mml9f" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.090207 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.234396 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-config\") pod \"9e62dafe-2ab3-467f-adda-d18766ddd450\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.234485 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-dns-svc\") pod \"9e62dafe-2ab3-467f-adda-d18766ddd450\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.250852 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dqvc\" (UniqueName: \"kubernetes.io/projected/9e62dafe-2ab3-467f-adda-d18766ddd450-kube-api-access-6dqvc\") pod \"9e62dafe-2ab3-467f-adda-d18766ddd450\" (UID: \"9e62dafe-2ab3-467f-adda-d18766ddd450\") " Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.255154 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e62dafe-2ab3-467f-adda-d18766ddd450-kube-api-access-6dqvc" (OuterVolumeSpecName: "kube-api-access-6dqvc") pod "9e62dafe-2ab3-467f-adda-d18766ddd450" (UID: "9e62dafe-2ab3-467f-adda-d18766ddd450"). InnerVolumeSpecName "kube-api-access-6dqvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.290449 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e62dafe-2ab3-467f-adda-d18766ddd450" (UID: "9e62dafe-2ab3-467f-adda-d18766ddd450"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.319829 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-config" (OuterVolumeSpecName: "config") pod "9e62dafe-2ab3-467f-adda-d18766ddd450" (UID: "9e62dafe-2ab3-467f-adda-d18766ddd450"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.353766 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.353800 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e62dafe-2ab3-467f-adda-d18766ddd450-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:27 crc kubenswrapper[4788]: I1010 15:04:27.353809 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dqvc\" (UniqueName: \"kubernetes.io/projected/9e62dafe-2ab3-467f-adda-d18766ddd450-kube-api-access-6dqvc\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.033023 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" event={"ID":"9e62dafe-2ab3-467f-adda-d18766ddd450","Type":"ContainerDied","Data":"314fad5b5d8e18c796b7751ac1bc02f583c7e0e0acf335937275d9ff287e674a"} Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.033354 4788 scope.go:117] "RemoveContainer" containerID="2790fb9ce8289471d956a6a24c47b65328923fed0f5f4a1c6f45e886c788dcf4" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.033094 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-r4mvh" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.059859 4788 scope.go:117] "RemoveContainer" containerID="10d08519d4c7317c152b6029114795663f2138c64e9ebdd0c519113d68276b30" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.065457 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-r4mvh"] Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.073428 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-r4mvh"] Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.246796 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e62dafe-2ab3-467f-adda-d18766ddd450" path="/var/lib/kubelet/pods/9e62dafe-2ab3-467f-adda-d18766ddd450/volumes" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.693678 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-gxtng"] Oct 10 15:04:28 crc kubenswrapper[4788]: E1010 15:04:28.694100 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9f621aa-8944-4699-b706-53c84c2e9fe5" containerName="mariadb-database-create" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.694120 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9f621aa-8944-4699-b706-53c84c2e9fe5" containerName="mariadb-database-create" Oct 10 15:04:28 crc kubenswrapper[4788]: E1010 15:04:28.694165 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e62dafe-2ab3-467f-adda-d18766ddd450" containerName="init" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.694175 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e62dafe-2ab3-467f-adda-d18766ddd450" containerName="init" Oct 10 15:04:28 crc kubenswrapper[4788]: E1010 15:04:28.694199 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e62dafe-2ab3-467f-adda-d18766ddd450" containerName="dnsmasq-dns" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.694209 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e62dafe-2ab3-467f-adda-d18766ddd450" containerName="dnsmasq-dns" Oct 10 15:04:28 crc kubenswrapper[4788]: E1010 15:04:28.694224 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12ed2f7-2d5e-4a57-bd17-59537de8065a" containerName="mariadb-database-create" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.694233 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12ed2f7-2d5e-4a57-bd17-59537de8065a" containerName="mariadb-database-create" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.694431 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9f621aa-8944-4699-b706-53c84c2e9fe5" containerName="mariadb-database-create" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.694449 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e62dafe-2ab3-467f-adda-d18766ddd450" containerName="dnsmasq-dns" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.694459 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12ed2f7-2d5e-4a57-bd17-59537de8065a" containerName="mariadb-database-create" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.695266 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gxtng" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.719371 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-gxtng"] Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.881083 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrd6g\" (UniqueName: \"kubernetes.io/projected/7932fad1-fc89-4cde-86d4-210a1121424d-kube-api-access-jrd6g\") pod \"glance-db-create-gxtng\" (UID: \"7932fad1-fc89-4cde-86d4-210a1121424d\") " pod="openstack/glance-db-create-gxtng" Oct 10 15:04:28 crc kubenswrapper[4788]: I1010 15:04:28.983062 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrd6g\" (UniqueName: \"kubernetes.io/projected/7932fad1-fc89-4cde-86d4-210a1121424d-kube-api-access-jrd6g\") pod \"glance-db-create-gxtng\" (UID: \"7932fad1-fc89-4cde-86d4-210a1121424d\") " pod="openstack/glance-db-create-gxtng" Oct 10 15:04:29 crc kubenswrapper[4788]: I1010 15:04:29.018817 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrd6g\" (UniqueName: \"kubernetes.io/projected/7932fad1-fc89-4cde-86d4-210a1121424d-kube-api-access-jrd6g\") pod \"glance-db-create-gxtng\" (UID: \"7932fad1-fc89-4cde-86d4-210a1121424d\") " pod="openstack/glance-db-create-gxtng" Oct 10 15:04:29 crc kubenswrapper[4788]: I1010 15:04:29.019581 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gxtng" Oct 10 15:04:29 crc kubenswrapper[4788]: I1010 15:04:29.496900 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-gxtng"] Oct 10 15:04:29 crc kubenswrapper[4788]: W1010 15:04:29.508748 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7932fad1_fc89_4cde_86d4_210a1121424d.slice/crio-d15fc6a406e421e0c0507a92b27719d7139f645865e436d989b532eca471b624 WatchSource:0}: Error finding container d15fc6a406e421e0c0507a92b27719d7139f645865e436d989b532eca471b624: Status 404 returned error can't find the container with id d15fc6a406e421e0c0507a92b27719d7139f645865e436d989b532eca471b624 Oct 10 15:04:30 crc kubenswrapper[4788]: I1010 15:04:30.052971 4788 generic.go:334] "Generic (PLEG): container finished" podID="7932fad1-fc89-4cde-86d4-210a1121424d" containerID="87b5409f5a03eb92c52a01f84e674066fa474f0acba5914786ae0ae6c9f1504d" exitCode=0 Oct 10 15:04:30 crc kubenswrapper[4788]: I1010 15:04:30.053026 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gxtng" event={"ID":"7932fad1-fc89-4cde-86d4-210a1121424d","Type":"ContainerDied","Data":"87b5409f5a03eb92c52a01f84e674066fa474f0acba5914786ae0ae6c9f1504d"} Oct 10 15:04:30 crc kubenswrapper[4788]: I1010 15:04:30.053559 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gxtng" event={"ID":"7932fad1-fc89-4cde-86d4-210a1121424d","Type":"ContainerStarted","Data":"d15fc6a406e421e0c0507a92b27719d7139f645865e436d989b532eca471b624"} Oct 10 15:04:30 crc kubenswrapper[4788]: I1010 15:04:30.055998 4788 generic.go:334] "Generic (PLEG): container finished" podID="bb647ff7-3bd1-4325-a6a6-916a6f134fea" containerID="3af2beb51561d3ef0612a1db993a775db975046929eb4d1f3edfa355eb9cecc3" exitCode=0 Oct 10 15:04:30 crc kubenswrapper[4788]: I1010 15:04:30.056046 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fr4km" event={"ID":"bb647ff7-3bd1-4325-a6a6-916a6f134fea","Type":"ContainerDied","Data":"3af2beb51561d3ef0612a1db993a775db975046929eb4d1f3edfa355eb9cecc3"} Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.668103 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.670119 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gxtng" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.857311 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bb647ff7-3bd1-4325-a6a6-916a6f134fea" (UID: "bb647ff7-3bd1-4325-a6a6-916a6f134fea"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.855836 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-ring-data-devices\") pod \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.857554 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrd6g\" (UniqueName: \"kubernetes.io/projected/7932fad1-fc89-4cde-86d4-210a1121424d-kube-api-access-jrd6g\") pod \"7932fad1-fc89-4cde-86d4-210a1121424d\" (UID: \"7932fad1-fc89-4cde-86d4-210a1121424d\") " Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.858865 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-combined-ca-bundle\") pod \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.858936 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4h6c\" (UniqueName: \"kubernetes.io/projected/bb647ff7-3bd1-4325-a6a6-916a6f134fea-kube-api-access-x4h6c\") pod \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.859011 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-swiftconf\") pod \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.859046 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-dispersionconf\") pod \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.859088 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb647ff7-3bd1-4325-a6a6-916a6f134fea-etc-swift\") pod \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.859442 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-scripts\") pod \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\" (UID: \"bb647ff7-3bd1-4325-a6a6-916a6f134fea\") " Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.860188 4788 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.860950 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb647ff7-3bd1-4325-a6a6-916a6f134fea-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bb647ff7-3bd1-4325-a6a6-916a6f134fea" (UID: "bb647ff7-3bd1-4325-a6a6-916a6f134fea"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.868581 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7932fad1-fc89-4cde-86d4-210a1121424d-kube-api-access-jrd6g" (OuterVolumeSpecName: "kube-api-access-jrd6g") pod "7932fad1-fc89-4cde-86d4-210a1121424d" (UID: "7932fad1-fc89-4cde-86d4-210a1121424d"). InnerVolumeSpecName "kube-api-access-jrd6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.872551 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb647ff7-3bd1-4325-a6a6-916a6f134fea-kube-api-access-x4h6c" (OuterVolumeSpecName: "kube-api-access-x4h6c") pod "bb647ff7-3bd1-4325-a6a6-916a6f134fea" (UID: "bb647ff7-3bd1-4325-a6a6-916a6f134fea"). InnerVolumeSpecName "kube-api-access-x4h6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.875693 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.887507 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bb647ff7-3bd1-4325-a6a6-916a6f134fea" (UID: "bb647ff7-3bd1-4325-a6a6-916a6f134fea"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.892422 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bb647ff7-3bd1-4325-a6a6-916a6f134fea" (UID: "bb647ff7-3bd1-4325-a6a6-916a6f134fea"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.908777 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-scripts" (OuterVolumeSpecName: "scripts") pod "bb647ff7-3bd1-4325-a6a6-916a6f134fea" (UID: "bb647ff7-3bd1-4325-a6a6-916a6f134fea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.909978 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb647ff7-3bd1-4325-a6a6-916a6f134fea" (UID: "bb647ff7-3bd1-4325-a6a6-916a6f134fea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.961969 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrd6g\" (UniqueName: \"kubernetes.io/projected/7932fad1-fc89-4cde-86d4-210a1121424d-kube-api-access-jrd6g\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.962023 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.962037 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4h6c\" (UniqueName: \"kubernetes.io/projected/bb647ff7-3bd1-4325-a6a6-916a6f134fea-kube-api-access-x4h6c\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.962053 4788 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.962073 4788 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb647ff7-3bd1-4325-a6a6-916a6f134fea-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.962084 4788 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb647ff7-3bd1-4325-a6a6-916a6f134fea-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:31 crc kubenswrapper[4788]: I1010 15:04:31.962098 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb647ff7-3bd1-4325-a6a6-916a6f134fea-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.082267 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gxtng" event={"ID":"7932fad1-fc89-4cde-86d4-210a1121424d","Type":"ContainerDied","Data":"d15fc6a406e421e0c0507a92b27719d7139f645865e436d989b532eca471b624"} Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.082339 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15fc6a406e421e0c0507a92b27719d7139f645865e436d989b532eca471b624" Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.082310 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gxtng" Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.085129 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fr4km" event={"ID":"bb647ff7-3bd1-4325-a6a6-916a6f134fea","Type":"ContainerDied","Data":"833b0c2dded97b1a0bcf5b98261e897597ff700f911b865b6a640d1223a007dc"} Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.085176 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="833b0c2dded97b1a0bcf5b98261e897597ff700f911b865b6a640d1223a007dc" Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.085375 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fr4km" Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.678942 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.687468 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"swift-storage-0\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " pod="openstack/swift-storage-0" Oct 10 15:04:32 crc kubenswrapper[4788]: I1010 15:04:32.984700 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.227238 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-56b0-account-create-wqxf2"] Oct 10 15:04:33 crc kubenswrapper[4788]: E1010 15:04:33.227886 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7932fad1-fc89-4cde-86d4-210a1121424d" containerName="mariadb-database-create" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.227910 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7932fad1-fc89-4cde-86d4-210a1121424d" containerName="mariadb-database-create" Oct 10 15:04:33 crc kubenswrapper[4788]: E1010 15:04:33.227933 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb647ff7-3bd1-4325-a6a6-916a6f134fea" containerName="swift-ring-rebalance" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.227941 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb647ff7-3bd1-4325-a6a6-916a6f134fea" containerName="swift-ring-rebalance" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.228232 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb647ff7-3bd1-4325-a6a6-916a6f134fea" containerName="swift-ring-rebalance" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.228265 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7932fad1-fc89-4cde-86d4-210a1121424d" containerName="mariadb-database-create" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.229226 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56b0-account-create-wqxf2" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.231971 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.238305 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56b0-account-create-wqxf2"] Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.292303 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsztk\" (UniqueName: \"kubernetes.io/projected/5dc79ccd-3195-4403-a56a-01c7ff0984a9-kube-api-access-wsztk\") pod \"keystone-56b0-account-create-wqxf2\" (UID: \"5dc79ccd-3195-4403-a56a-01c7ff0984a9\") " pod="openstack/keystone-56b0-account-create-wqxf2" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.395857 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsztk\" (UniqueName: \"kubernetes.io/projected/5dc79ccd-3195-4403-a56a-01c7ff0984a9-kube-api-access-wsztk\") pod \"keystone-56b0-account-create-wqxf2\" (UID: \"5dc79ccd-3195-4403-a56a-01c7ff0984a9\") " pod="openstack/keystone-56b0-account-create-wqxf2" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.428389 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsztk\" (UniqueName: \"kubernetes.io/projected/5dc79ccd-3195-4403-a56a-01c7ff0984a9-kube-api-access-wsztk\") pod \"keystone-56b0-account-create-wqxf2\" (UID: \"5dc79ccd-3195-4403-a56a-01c7ff0984a9\") " pod="openstack/keystone-56b0-account-create-wqxf2" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.558004 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56b0-account-create-wqxf2" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.559940 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4ba5-account-create-jhhcp"] Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.561446 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4ba5-account-create-jhhcp" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.564121 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.570226 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4ba5-account-create-jhhcp"] Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.611010 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm72h\" (UniqueName: \"kubernetes.io/projected/d89473fa-b587-4035-a586-80f51fa65786-kube-api-access-qm72h\") pod \"placement-4ba5-account-create-jhhcp\" (UID: \"d89473fa-b587-4035-a586-80f51fa65786\") " pod="openstack/placement-4ba5-account-create-jhhcp" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.664474 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 10 15:04:33 crc kubenswrapper[4788]: W1010 15:04:33.680598 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedada8e6_c183_4ef8_bcd5_f5c617e7615e.slice/crio-a2b03249af073277654e5468956d1eebd61d5943a330291a73eb30edd0ffbd6e WatchSource:0}: Error finding container a2b03249af073277654e5468956d1eebd61d5943a330291a73eb30edd0ffbd6e: Status 404 returned error can't find the container with id a2b03249af073277654e5468956d1eebd61d5943a330291a73eb30edd0ffbd6e Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.713756 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm72h\" (UniqueName: \"kubernetes.io/projected/d89473fa-b587-4035-a586-80f51fa65786-kube-api-access-qm72h\") pod \"placement-4ba5-account-create-jhhcp\" (UID: \"d89473fa-b587-4035-a586-80f51fa65786\") " pod="openstack/placement-4ba5-account-create-jhhcp" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.737464 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm72h\" (UniqueName: \"kubernetes.io/projected/d89473fa-b587-4035-a586-80f51fa65786-kube-api-access-qm72h\") pod \"placement-4ba5-account-create-jhhcp\" (UID: \"d89473fa-b587-4035-a586-80f51fa65786\") " pod="openstack/placement-4ba5-account-create-jhhcp" Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.837707 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-n84gv" podUID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" containerName="ovn-controller" probeResult="failure" output=< Oct 10 15:04:33 crc kubenswrapper[4788]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 10 15:04:33 crc kubenswrapper[4788]: > Oct 10 15:04:33 crc kubenswrapper[4788]: I1010 15:04:33.967896 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4ba5-account-create-jhhcp" Oct 10 15:04:34 crc kubenswrapper[4788]: I1010 15:04:34.051263 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56b0-account-create-wqxf2"] Oct 10 15:04:34 crc kubenswrapper[4788]: W1010 15:04:34.062023 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc79ccd_3195_4403_a56a_01c7ff0984a9.slice/crio-73fa2fff9910b8b1b9664247e6b0b6525c3c333e6e4351479bbfe9cfb5dfab36 WatchSource:0}: Error finding container 73fa2fff9910b8b1b9664247e6b0b6525c3c333e6e4351479bbfe9cfb5dfab36: Status 404 returned error can't find the container with id 73fa2fff9910b8b1b9664247e6b0b6525c3c333e6e4351479bbfe9cfb5dfab36 Oct 10 15:04:34 crc kubenswrapper[4788]: I1010 15:04:34.122343 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"a2b03249af073277654e5468956d1eebd61d5943a330291a73eb30edd0ffbd6e"} Oct 10 15:04:34 crc kubenswrapper[4788]: I1010 15:04:34.124722 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56b0-account-create-wqxf2" event={"ID":"5dc79ccd-3195-4403-a56a-01c7ff0984a9","Type":"ContainerStarted","Data":"73fa2fff9910b8b1b9664247e6b0b6525c3c333e6e4351479bbfe9cfb5dfab36"} Oct 10 15:04:34 crc kubenswrapper[4788]: I1010 15:04:34.436328 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4ba5-account-create-jhhcp"] Oct 10 15:04:35 crc kubenswrapper[4788]: I1010 15:04:35.144870 4788 generic.go:334] "Generic (PLEG): container finished" podID="5dc79ccd-3195-4403-a56a-01c7ff0984a9" containerID="5a5958beeb484fa10133ebb3bdf2fb0ffb0366c0052b30e8639799c67e07fcc9" exitCode=0 Oct 10 15:04:35 crc kubenswrapper[4788]: I1010 15:04:35.144985 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56b0-account-create-wqxf2" event={"ID":"5dc79ccd-3195-4403-a56a-01c7ff0984a9","Type":"ContainerDied","Data":"5a5958beeb484fa10133ebb3bdf2fb0ffb0366c0052b30e8639799c67e07fcc9"} Oct 10 15:04:35 crc kubenswrapper[4788]: I1010 15:04:35.146991 4788 generic.go:334] "Generic (PLEG): container finished" podID="d89473fa-b587-4035-a586-80f51fa65786" containerID="0e206d9712b2fcf1d859caed9a81a00ff94cd840d0e81d1c761455259fa5b1b2" exitCode=0 Oct 10 15:04:35 crc kubenswrapper[4788]: I1010 15:04:35.147057 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4ba5-account-create-jhhcp" event={"ID":"d89473fa-b587-4035-a586-80f51fa65786","Type":"ContainerDied","Data":"0e206d9712b2fcf1d859caed9a81a00ff94cd840d0e81d1c761455259fa5b1b2"} Oct 10 15:04:35 crc kubenswrapper[4788]: I1010 15:04:35.147098 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4ba5-account-create-jhhcp" event={"ID":"d89473fa-b587-4035-a586-80f51fa65786","Type":"ContainerStarted","Data":"08f005234bf579e637fa6ce5350f5eb10e855d6e35ac06ba3afba4f73c4332bc"} Oct 10 15:04:35 crc kubenswrapper[4788]: I1010 15:04:35.148550 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"3c7dd5fa50a9cff710f02fddbf4c3623aa2fc790f2bae6d263715c6a2a0be29a"} Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.168423 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"3ffc6ad965d6e1a531e493d87c29df6c3edcf3b174067bab2eec1c073223fd6c"} Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.171369 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"c68f1e8fc4064c0e04c7dd51691e00ff828f26e781caa02ae72f7ca08e509516"} Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.171445 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"bf79745568579090730afa53a1af2d611537375a823410f2aefdb5bb4ae1de4c"} Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.618659 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4ba5-account-create-jhhcp" Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.624563 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56b0-account-create-wqxf2" Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.792230 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm72h\" (UniqueName: \"kubernetes.io/projected/d89473fa-b587-4035-a586-80f51fa65786-kube-api-access-qm72h\") pod \"d89473fa-b587-4035-a586-80f51fa65786\" (UID: \"d89473fa-b587-4035-a586-80f51fa65786\") " Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.792806 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsztk\" (UniqueName: \"kubernetes.io/projected/5dc79ccd-3195-4403-a56a-01c7ff0984a9-kube-api-access-wsztk\") pod \"5dc79ccd-3195-4403-a56a-01c7ff0984a9\" (UID: \"5dc79ccd-3195-4403-a56a-01c7ff0984a9\") " Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.799670 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d89473fa-b587-4035-a586-80f51fa65786-kube-api-access-qm72h" (OuterVolumeSpecName: "kube-api-access-qm72h") pod "d89473fa-b587-4035-a586-80f51fa65786" (UID: "d89473fa-b587-4035-a586-80f51fa65786"). InnerVolumeSpecName "kube-api-access-qm72h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.800266 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc79ccd-3195-4403-a56a-01c7ff0984a9-kube-api-access-wsztk" (OuterVolumeSpecName: "kube-api-access-wsztk") pod "5dc79ccd-3195-4403-a56a-01c7ff0984a9" (UID: "5dc79ccd-3195-4403-a56a-01c7ff0984a9"). InnerVolumeSpecName "kube-api-access-wsztk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.895056 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsztk\" (UniqueName: \"kubernetes.io/projected/5dc79ccd-3195-4403-a56a-01c7ff0984a9-kube-api-access-wsztk\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:36 crc kubenswrapper[4788]: I1010 15:04:36.895093 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm72h\" (UniqueName: \"kubernetes.io/projected/d89473fa-b587-4035-a586-80f51fa65786-kube-api-access-qm72h\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:37 crc kubenswrapper[4788]: I1010 15:04:37.179762 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56b0-account-create-wqxf2" event={"ID":"5dc79ccd-3195-4403-a56a-01c7ff0984a9","Type":"ContainerDied","Data":"73fa2fff9910b8b1b9664247e6b0b6525c3c333e6e4351479bbfe9cfb5dfab36"} Oct 10 15:04:37 crc kubenswrapper[4788]: I1010 15:04:37.179847 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73fa2fff9910b8b1b9664247e6b0b6525c3c333e6e4351479bbfe9cfb5dfab36" Oct 10 15:04:37 crc kubenswrapper[4788]: I1010 15:04:37.179956 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56b0-account-create-wqxf2" Oct 10 15:04:37 crc kubenswrapper[4788]: I1010 15:04:37.181977 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4ba5-account-create-jhhcp" event={"ID":"d89473fa-b587-4035-a586-80f51fa65786","Type":"ContainerDied","Data":"08f005234bf579e637fa6ce5350f5eb10e855d6e35ac06ba3afba4f73c4332bc"} Oct 10 15:04:37 crc kubenswrapper[4788]: I1010 15:04:37.182108 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08f005234bf579e637fa6ce5350f5eb10e855d6e35ac06ba3afba4f73c4332bc" Oct 10 15:04:37 crc kubenswrapper[4788]: I1010 15:04:37.182007 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4ba5-account-create-jhhcp" Oct 10 15:04:37 crc kubenswrapper[4788]: I1010 15:04:37.183513 4788 generic.go:334] "Generic (PLEG): container finished" podID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerID="a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290" exitCode=0 Oct 10 15:04:37 crc kubenswrapper[4788]: I1010 15:04:37.183563 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dcf4b178-75ee-47fb-81e4-e6b277eb1790","Type":"ContainerDied","Data":"a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290"} Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.205801 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"642b90028a4eb24c71624969b7de73f0a0d672a8124912cb4fa0a195211e072e"} Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.220941 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dcf4b178-75ee-47fb-81e4-e6b277eb1790","Type":"ContainerStarted","Data":"c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251"} Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.221210 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.224978 4788 generic.go:334] "Generic (PLEG): container finished" podID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerID="54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45" exitCode=0 Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.225014 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c","Type":"ContainerDied","Data":"54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45"} Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.268712 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=52.189282335 podStartE2EDuration="1m0.268686611s" podCreationTimestamp="2025-10-10 15:03:38 +0000 UTC" firstStartedPulling="2025-10-10 15:03:54.470987357 +0000 UTC m=+1136.920702905" lastFinishedPulling="2025-10-10 15:04:02.550391633 +0000 UTC m=+1145.000107181" observedRunningTime="2025-10-10 15:04:38.262271048 +0000 UTC m=+1180.711986606" watchObservedRunningTime="2025-10-10 15:04:38.268686611 +0000 UTC m=+1180.718402169" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.841560 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-9145-account-create-vxxpk"] Oct 10 15:04:38 crc kubenswrapper[4788]: E1010 15:04:38.842075 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc79ccd-3195-4403-a56a-01c7ff0984a9" containerName="mariadb-account-create" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.842091 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc79ccd-3195-4403-a56a-01c7ff0984a9" containerName="mariadb-account-create" Oct 10 15:04:38 crc kubenswrapper[4788]: E1010 15:04:38.842111 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d89473fa-b587-4035-a586-80f51fa65786" containerName="mariadb-account-create" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.842118 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d89473fa-b587-4035-a586-80f51fa65786" containerName="mariadb-account-create" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.842334 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d89473fa-b587-4035-a586-80f51fa65786" containerName="mariadb-account-create" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.842367 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc79ccd-3195-4403-a56a-01c7ff0984a9" containerName="mariadb-account-create" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.843031 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9145-account-create-vxxpk" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.846744 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.853052 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9145-account-create-vxxpk"] Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.879436 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-n84gv" podUID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" containerName="ovn-controller" probeResult="failure" output=< Oct 10 15:04:38 crc kubenswrapper[4788]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 10 15:04:38 crc kubenswrapper[4788]: > Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.925332 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.929682 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:04:38 crc kubenswrapper[4788]: I1010 15:04:38.933731 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvsj5\" (UniqueName: \"kubernetes.io/projected/aafa0506-d825-478a-9dc8-79f11ad28017-kube-api-access-wvsj5\") pod \"glance-9145-account-create-vxxpk\" (UID: \"aafa0506-d825-478a-9dc8-79f11ad28017\") " pod="openstack/glance-9145-account-create-vxxpk" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.035683 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvsj5\" (UniqueName: \"kubernetes.io/projected/aafa0506-d825-478a-9dc8-79f11ad28017-kube-api-access-wvsj5\") pod \"glance-9145-account-create-vxxpk\" (UID: \"aafa0506-d825-478a-9dc8-79f11ad28017\") " pod="openstack/glance-9145-account-create-vxxpk" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.079317 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvsj5\" (UniqueName: \"kubernetes.io/projected/aafa0506-d825-478a-9dc8-79f11ad28017-kube-api-access-wvsj5\") pod \"glance-9145-account-create-vxxpk\" (UID: \"aafa0506-d825-478a-9dc8-79f11ad28017\") " pod="openstack/glance-9145-account-create-vxxpk" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.170090 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-n84gv-config-c5zt2"] Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.171592 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.174683 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.180467 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9145-account-create-vxxpk" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.191229 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-n84gv-config-c5zt2"] Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.238478 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-log-ovn\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.238572 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-scripts\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.238608 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run-ovn\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.238675 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.238706 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-additional-scripts\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.238735 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f2ks\" (UniqueName: \"kubernetes.io/projected/6a4335e2-ddc0-4147-8fba-4bf0698c6843-kube-api-access-7f2ks\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.276323 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c","Type":"ContainerStarted","Data":"a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab"} Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.276850 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.316406 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.879452157 podStartE2EDuration="1m0.316388237s" podCreationTimestamp="2025-10-10 15:03:39 +0000 UTC" firstStartedPulling="2025-10-10 15:03:54.476478315 +0000 UTC m=+1136.926193863" lastFinishedPulling="2025-10-10 15:04:02.913414405 +0000 UTC m=+1145.363129943" observedRunningTime="2025-10-10 15:04:39.30986857 +0000 UTC m=+1181.759584118" watchObservedRunningTime="2025-10-10 15:04:39.316388237 +0000 UTC m=+1181.766103785" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.322383 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"7e17d3abb98cf36468ce697828f20b0a33f4bdac07da3b2f7a9793d696601d10"} Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.322446 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"4e4b1fea33830faff4707c7c32b35feb21cbafff038fad525b1d34bd1ddc3463"} Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.322460 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"da1d35d43d8e68922f14f6c617628576709461d2e0a8268d451f4cc8d694b90c"} Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.343248 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.343345 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-additional-scripts\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.343389 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f2ks\" (UniqueName: \"kubernetes.io/projected/6a4335e2-ddc0-4147-8fba-4bf0698c6843-kube-api-access-7f2ks\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.343429 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-log-ovn\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.343507 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-scripts\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.343534 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run-ovn\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.343995 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run-ovn\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.343994 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.344084 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-log-ovn\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.344785 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-additional-scripts\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.347037 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-scripts\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.385600 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f2ks\" (UniqueName: \"kubernetes.io/projected/6a4335e2-ddc0-4147-8fba-4bf0698c6843-kube-api-access-7f2ks\") pod \"ovn-controller-n84gv-config-c5zt2\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.587379 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:39 crc kubenswrapper[4788]: I1010 15:04:39.794614 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9145-account-create-vxxpk"] Oct 10 15:04:40 crc kubenswrapper[4788]: I1010 15:04:40.067911 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-n84gv-config-c5zt2"] Oct 10 15:04:40 crc kubenswrapper[4788]: W1010 15:04:40.090201 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a4335e2_ddc0_4147_8fba_4bf0698c6843.slice/crio-f50419cfd0c5675ad08621d544e3132c03279dd6c568ab672852c86ae905c827 WatchSource:0}: Error finding container f50419cfd0c5675ad08621d544e3132c03279dd6c568ab672852c86ae905c827: Status 404 returned error can't find the container with id f50419cfd0c5675ad08621d544e3132c03279dd6c568ab672852c86ae905c827 Oct 10 15:04:40 crc kubenswrapper[4788]: I1010 15:04:40.342848 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n84gv-config-c5zt2" event={"ID":"6a4335e2-ddc0-4147-8fba-4bf0698c6843","Type":"ContainerStarted","Data":"f50419cfd0c5675ad08621d544e3132c03279dd6c568ab672852c86ae905c827"} Oct 10 15:04:40 crc kubenswrapper[4788]: I1010 15:04:40.347013 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9145-account-create-vxxpk" event={"ID":"aafa0506-d825-478a-9dc8-79f11ad28017","Type":"ContainerStarted","Data":"f0a4c089fd7d0b89314e500a54e9e888fb4099f6bd2b5736fa18d7a297dc7323"} Oct 10 15:04:40 crc kubenswrapper[4788]: I1010 15:04:40.347089 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9145-account-create-vxxpk" event={"ID":"aafa0506-d825-478a-9dc8-79f11ad28017","Type":"ContainerStarted","Data":"f5a71fae1914a32482af1ea256660098c78af6b937c0fd3e5b5affbfe1a19738"} Oct 10 15:04:40 crc kubenswrapper[4788]: I1010 15:04:40.365280 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"d37f41271c5034f6009e04dd6c3c22abb922d7b658aa8853581bf17fef626476"} Oct 10 15:04:40 crc kubenswrapper[4788]: I1010 15:04:40.369398 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-9145-account-create-vxxpk" podStartSLOduration=2.369379235 podStartE2EDuration="2.369379235s" podCreationTimestamp="2025-10-10 15:04:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:04:40.365709797 +0000 UTC m=+1182.815425345" watchObservedRunningTime="2025-10-10 15:04:40.369379235 +0000 UTC m=+1182.819094783" Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.372550 4788 generic.go:334] "Generic (PLEG): container finished" podID="6a4335e2-ddc0-4147-8fba-4bf0698c6843" containerID="081f9a04373c71c05b380d5f33c6e373a025a23595bb2ef7f84dd4e4b4b298fe" exitCode=0 Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.372758 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n84gv-config-c5zt2" event={"ID":"6a4335e2-ddc0-4147-8fba-4bf0698c6843","Type":"ContainerDied","Data":"081f9a04373c71c05b380d5f33c6e373a025a23595bb2ef7f84dd4e4b4b298fe"} Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.377768 4788 generic.go:334] "Generic (PLEG): container finished" podID="aafa0506-d825-478a-9dc8-79f11ad28017" containerID="f0a4c089fd7d0b89314e500a54e9e888fb4099f6bd2b5736fa18d7a297dc7323" exitCode=0 Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.377920 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9145-account-create-vxxpk" event={"ID":"aafa0506-d825-478a-9dc8-79f11ad28017","Type":"ContainerDied","Data":"f0a4c089fd7d0b89314e500a54e9e888fb4099f6bd2b5736fa18d7a297dc7323"} Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.385752 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"e1fa0e72188924318b11ba0840f43b627695d374bb14fe291577b87bd9109b71"} Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.385802 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"21748426974e63b0540c88af385f3f29ae489df2ec5f695b422d31559e86f185"} Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.385814 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"397c0d1e240027ba5e81338ba3e8dbb1354f4d23cae97d7d881c24cc8254d588"} Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.385824 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"bb676839b319bffb1dc7ed545f7044cc6b58cf9440fc39529f2f530eaababd0f"} Oct 10 15:04:41 crc kubenswrapper[4788]: I1010 15:04:41.385833 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"292cc90a59ee5b84483d07d2af032969b7328c7cc913e39ba2daca604f9df39c"} Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.406095 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerStarted","Data":"ea921c62533da26ac37b94fc9c80bfc74b41e75cddc9adab1f4225855ea56a7d"} Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.447706 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.028316584 podStartE2EDuration="27.447680414s" podCreationTimestamp="2025-10-10 15:04:15 +0000 UTC" firstStartedPulling="2025-10-10 15:04:33.68535027 +0000 UTC m=+1176.135065818" lastFinishedPulling="2025-10-10 15:04:40.1047141 +0000 UTC m=+1182.554429648" observedRunningTime="2025-10-10 15:04:42.43934396 +0000 UTC m=+1184.889059508" watchObservedRunningTime="2025-10-10 15:04:42.447680414 +0000 UTC m=+1184.897395962" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.813428 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-4dn9v"] Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.815515 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.825730 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.837444 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-4dn9v"] Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.911320 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stcnb\" (UniqueName: \"kubernetes.io/projected/f654d8a6-43f1-46e7-b046-d1f52af77f48-kube-api-access-stcnb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.911373 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-config\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.911398 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.911473 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.911504 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.911542 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.916272 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9145-account-create-vxxpk" Oct 10 15:04:42 crc kubenswrapper[4788]: I1010 15:04:42.924188 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.013318 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run\") pod \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.013514 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f2ks\" (UniqueName: \"kubernetes.io/projected/6a4335e2-ddc0-4147-8fba-4bf0698c6843-kube-api-access-7f2ks\") pod \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.013553 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-scripts\") pod \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.013535 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run" (OuterVolumeSpecName: "var-run") pod "6a4335e2-ddc0-4147-8fba-4bf0698c6843" (UID: "6a4335e2-ddc0-4147-8fba-4bf0698c6843"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.013611 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvsj5\" (UniqueName: \"kubernetes.io/projected/aafa0506-d825-478a-9dc8-79f11ad28017-kube-api-access-wvsj5\") pod \"aafa0506-d825-478a-9dc8-79f11ad28017\" (UID: \"aafa0506-d825-478a-9dc8-79f11ad28017\") " Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.013737 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run-ovn\") pod \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.013815 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-additional-scripts\") pod \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.013960 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-log-ovn\") pod \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\" (UID: \"6a4335e2-ddc0-4147-8fba-4bf0698c6843\") " Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.014127 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6a4335e2-ddc0-4147-8fba-4bf0698c6843" (UID: "6a4335e2-ddc0-4147-8fba-4bf0698c6843"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.014361 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6a4335e2-ddc0-4147-8fba-4bf0698c6843" (UID: "6a4335e2-ddc0-4147-8fba-4bf0698c6843"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.014517 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stcnb\" (UniqueName: \"kubernetes.io/projected/f654d8a6-43f1-46e7-b046-d1f52af77f48-kube-api-access-stcnb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.014601 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-config\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.014655 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.015202 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.015271 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.015349 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.015522 4788 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.015539 4788 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.015551 4788 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a4335e2-ddc0-4147-8fba-4bf0698c6843-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.015555 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6a4335e2-ddc0-4147-8fba-4bf0698c6843" (UID: "6a4335e2-ddc0-4147-8fba-4bf0698c6843"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.015890 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-scripts" (OuterVolumeSpecName: "scripts") pod "6a4335e2-ddc0-4147-8fba-4bf0698c6843" (UID: "6a4335e2-ddc0-4147-8fba-4bf0698c6843"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.016399 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.016416 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-config\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.016657 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.016737 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.016898 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.021909 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a4335e2-ddc0-4147-8fba-4bf0698c6843-kube-api-access-7f2ks" (OuterVolumeSpecName: "kube-api-access-7f2ks") pod "6a4335e2-ddc0-4147-8fba-4bf0698c6843" (UID: "6a4335e2-ddc0-4147-8fba-4bf0698c6843"). InnerVolumeSpecName "kube-api-access-7f2ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.028117 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aafa0506-d825-478a-9dc8-79f11ad28017-kube-api-access-wvsj5" (OuterVolumeSpecName: "kube-api-access-wvsj5") pod "aafa0506-d825-478a-9dc8-79f11ad28017" (UID: "aafa0506-d825-478a-9dc8-79f11ad28017"). InnerVolumeSpecName "kube-api-access-wvsj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.036159 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stcnb\" (UniqueName: \"kubernetes.io/projected/f654d8a6-43f1-46e7-b046-d1f52af77f48-kube-api-access-stcnb\") pod \"dnsmasq-dns-77585f5f8c-4dn9v\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.116951 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f2ks\" (UniqueName: \"kubernetes.io/projected/6a4335e2-ddc0-4147-8fba-4bf0698c6843-kube-api-access-7f2ks\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.116993 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.117004 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvsj5\" (UniqueName: \"kubernetes.io/projected/aafa0506-d825-478a-9dc8-79f11ad28017-kube-api-access-wvsj5\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.117016 4788 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a4335e2-ddc0-4147-8fba-4bf0698c6843-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.210811 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.429801 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9145-account-create-vxxpk" event={"ID":"aafa0506-d825-478a-9dc8-79f11ad28017","Type":"ContainerDied","Data":"f5a71fae1914a32482af1ea256660098c78af6b937c0fd3e5b5affbfe1a19738"} Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.429873 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5a71fae1914a32482af1ea256660098c78af6b937c0fd3e5b5affbfe1a19738" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.429952 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9145-account-create-vxxpk" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.433817 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n84gv-config-c5zt2" event={"ID":"6a4335e2-ddc0-4147-8fba-4bf0698c6843","Type":"ContainerDied","Data":"f50419cfd0c5675ad08621d544e3132c03279dd6c568ab672852c86ae905c827"} Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.433854 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n84gv-config-c5zt2" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.433865 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f50419cfd0c5675ad08621d544e3132c03279dd6c568ab672852c86ae905c827" Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.688636 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-4dn9v"] Oct 10 15:04:43 crc kubenswrapper[4788]: W1010 15:04:43.698750 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf654d8a6_43f1_46e7_b046_d1f52af77f48.slice/crio-2aa2049e9dc7d9707bc9a774f2d300b98e84c8005f4cf85d741f36df4be51240 WatchSource:0}: Error finding container 2aa2049e9dc7d9707bc9a774f2d300b98e84c8005f4cf85d741f36df4be51240: Status 404 returned error can't find the container with id 2aa2049e9dc7d9707bc9a774f2d300b98e84c8005f4cf85d741f36df4be51240 Oct 10 15:04:43 crc kubenswrapper[4788]: I1010 15:04:43.851630 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-n84gv" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.017776 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-cpld9"] Oct 10 15:04:44 crc kubenswrapper[4788]: E1010 15:04:44.018742 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aafa0506-d825-478a-9dc8-79f11ad28017" containerName="mariadb-account-create" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.018763 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="aafa0506-d825-478a-9dc8-79f11ad28017" containerName="mariadb-account-create" Oct 10 15:04:44 crc kubenswrapper[4788]: E1010 15:04:44.018788 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a4335e2-ddc0-4147-8fba-4bf0698c6843" containerName="ovn-config" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.018798 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a4335e2-ddc0-4147-8fba-4bf0698c6843" containerName="ovn-config" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.019006 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a4335e2-ddc0-4147-8fba-4bf0698c6843" containerName="ovn-config" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.019041 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="aafa0506-d825-478a-9dc8-79f11ad28017" containerName="mariadb-account-create" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.019804 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.035484 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.035826 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-887mn" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.057580 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-cpld9"] Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.139489 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-config-data\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.139583 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whx96\" (UniqueName: \"kubernetes.io/projected/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-kube-api-access-whx96\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.139636 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-db-sync-config-data\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.139720 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-combined-ca-bundle\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.165268 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-n84gv-config-c5zt2"] Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.177682 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-n84gv-config-c5zt2"] Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.241272 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-combined-ca-bundle\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.241361 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-config-data\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.241408 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whx96\" (UniqueName: \"kubernetes.io/projected/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-kube-api-access-whx96\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.241454 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-db-sync-config-data\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.246718 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-db-sync-config-data\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.246753 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a4335e2-ddc0-4147-8fba-4bf0698c6843" path="/var/lib/kubelet/pods/6a4335e2-ddc0-4147-8fba-4bf0698c6843/volumes" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.249849 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-config-data\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.254878 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-combined-ca-bundle\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.265646 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whx96\" (UniqueName: \"kubernetes.io/projected/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-kube-api-access-whx96\") pod \"glance-db-sync-cpld9\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.340838 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-cpld9" Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.444939 4788 generic.go:334] "Generic (PLEG): container finished" podID="f654d8a6-43f1-46e7-b046-d1f52af77f48" containerID="4a6299b42981fdad04090c27687cca82ce0e81bf2af962149819d6c845b78f51" exitCode=0 Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.444977 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" event={"ID":"f654d8a6-43f1-46e7-b046-d1f52af77f48","Type":"ContainerDied","Data":"4a6299b42981fdad04090c27687cca82ce0e81bf2af962149819d6c845b78f51"} Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.445002 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" event={"ID":"f654d8a6-43f1-46e7-b046-d1f52af77f48","Type":"ContainerStarted","Data":"2aa2049e9dc7d9707bc9a774f2d300b98e84c8005f4cf85d741f36df4be51240"} Oct 10 15:04:44 crc kubenswrapper[4788]: I1010 15:04:44.985595 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-cpld9"] Oct 10 15:04:45 crc kubenswrapper[4788]: I1010 15:04:45.456531 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-cpld9" event={"ID":"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02","Type":"ContainerStarted","Data":"62b8035c071cdcbd5b5240c6bf7d7966d3ad283f4c95239ae9094fea421c691e"} Oct 10 15:04:45 crc kubenswrapper[4788]: I1010 15:04:45.459419 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" event={"ID":"f654d8a6-43f1-46e7-b046-d1f52af77f48","Type":"ContainerStarted","Data":"d603d0ab14e8d86d9580feb13b10db4d1d774aab6e488203ffbfdc111c16b2e8"} Oct 10 15:04:45 crc kubenswrapper[4788]: I1010 15:04:45.460872 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:45 crc kubenswrapper[4788]: I1010 15:04:45.490803 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" podStartSLOduration=3.490770682 podStartE2EDuration="3.490770682s" podCreationTimestamp="2025-10-10 15:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:04:45.483594258 +0000 UTC m=+1187.933309806" watchObservedRunningTime="2025-10-10 15:04:45.490770682 +0000 UTC m=+1187.940486240" Oct 10 15:04:50 crc kubenswrapper[4788]: I1010 15:04:50.569389 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 10 15:04:50 crc kubenswrapper[4788]: I1010 15:04:50.698727 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.030589 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-79446"] Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.032187 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-79446" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.052869 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-79446"] Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.076393 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxg72\" (UniqueName: \"kubernetes.io/projected/5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8-kube-api-access-qxg72\") pod \"cinder-db-create-79446\" (UID: \"5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8\") " pod="openstack/cinder-db-create-79446" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.127500 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-sclcr"] Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.129620 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sclcr" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.147787 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-sclcr"] Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.178510 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxg72\" (UniqueName: \"kubernetes.io/projected/5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8-kube-api-access-qxg72\") pod \"cinder-db-create-79446\" (UID: \"5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8\") " pod="openstack/cinder-db-create-79446" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.178615 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfjx2\" (UniqueName: \"kubernetes.io/projected/b5b098f0-5965-4ab3-a641-e301b2f599ed-kube-api-access-hfjx2\") pod \"barbican-db-create-sclcr\" (UID: \"b5b098f0-5965-4ab3-a641-e301b2f599ed\") " pod="openstack/barbican-db-create-sclcr" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.198543 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxg72\" (UniqueName: \"kubernetes.io/projected/5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8-kube-api-access-qxg72\") pod \"cinder-db-create-79446\" (UID: \"5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8\") " pod="openstack/cinder-db-create-79446" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.279950 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfjx2\" (UniqueName: \"kubernetes.io/projected/b5b098f0-5965-4ab3-a641-e301b2f599ed-kube-api-access-hfjx2\") pod \"barbican-db-create-sclcr\" (UID: \"b5b098f0-5965-4ab3-a641-e301b2f599ed\") " pod="openstack/barbican-db-create-sclcr" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.316447 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfjx2\" (UniqueName: \"kubernetes.io/projected/b5b098f0-5965-4ab3-a641-e301b2f599ed-kube-api-access-hfjx2\") pod \"barbican-db-create-sclcr\" (UID: \"b5b098f0-5965-4ab3-a641-e301b2f599ed\") " pod="openstack/barbican-db-create-sclcr" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.330242 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-c44p7"] Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.337422 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.355236 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-79446" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.356518 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.359582 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.361010 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wg8bv" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.365800 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.381224 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c44p7"] Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.382726 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-combined-ca-bundle\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.382838 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmxkq\" (UniqueName: \"kubernetes.io/projected/1063e9ab-7517-4a5e-8865-f7827c52245a-kube-api-access-gmxkq\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.382876 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-config-data\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.394271 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-cmdtx"] Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.395928 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cmdtx" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.405661 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-cmdtx"] Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.449975 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sclcr" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.498342 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmxkq\" (UniqueName: \"kubernetes.io/projected/1063e9ab-7517-4a5e-8865-f7827c52245a-kube-api-access-gmxkq\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.498385 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wdnb\" (UniqueName: \"kubernetes.io/projected/679c6797-8be3-49ee-ada2-c2ea883db467-kube-api-access-4wdnb\") pod \"neutron-db-create-cmdtx\" (UID: \"679c6797-8be3-49ee-ada2-c2ea883db467\") " pod="openstack/neutron-db-create-cmdtx" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.498429 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-config-data\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.498495 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-combined-ca-bundle\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.506864 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-combined-ca-bundle\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.511764 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-config-data\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.558962 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmxkq\" (UniqueName: \"kubernetes.io/projected/1063e9ab-7517-4a5e-8865-f7827c52245a-kube-api-access-gmxkq\") pod \"keystone-db-sync-c44p7\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.600189 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wdnb\" (UniqueName: \"kubernetes.io/projected/679c6797-8be3-49ee-ada2-c2ea883db467-kube-api-access-4wdnb\") pod \"neutron-db-create-cmdtx\" (UID: \"679c6797-8be3-49ee-ada2-c2ea883db467\") " pod="openstack/neutron-db-create-cmdtx" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.642001 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wdnb\" (UniqueName: \"kubernetes.io/projected/679c6797-8be3-49ee-ada2-c2ea883db467-kube-api-access-4wdnb\") pod \"neutron-db-create-cmdtx\" (UID: \"679c6797-8be3-49ee-ada2-c2ea883db467\") " pod="openstack/neutron-db-create-cmdtx" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.672864 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c44p7" Oct 10 15:04:51 crc kubenswrapper[4788]: I1010 15:04:51.721178 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cmdtx" Oct 10 15:04:53 crc kubenswrapper[4788]: I1010 15:04:53.212470 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:04:53 crc kubenswrapper[4788]: I1010 15:04:53.279198 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zjjvp"] Oct 10 15:04:53 crc kubenswrapper[4788]: I1010 15:04:53.279735 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-zjjvp" podUID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerName="dnsmasq-dns" containerID="cri-o://4ae05de9ee9dfbfd7f28da6127b6ad354efa3958ccfb741c52c9a125d68d2d1d" gracePeriod=10 Oct 10 15:04:53 crc kubenswrapper[4788]: I1010 15:04:53.558513 4788 generic.go:334] "Generic (PLEG): container finished" podID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerID="4ae05de9ee9dfbfd7f28da6127b6ad354efa3958ccfb741c52c9a125d68d2d1d" exitCode=0 Oct 10 15:04:53 crc kubenswrapper[4788]: I1010 15:04:53.558578 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zjjvp" event={"ID":"3f57f828-9ffd-4caf-872a-fc3fd1913df5","Type":"ContainerDied","Data":"4ae05de9ee9dfbfd7f28da6127b6ad354efa3958ccfb741c52c9a125d68d2d1d"} Oct 10 15:04:55 crc kubenswrapper[4788]: I1010 15:04:55.920184 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-zjjvp" podUID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.512330 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.594387 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-config\") pod \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.594420 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-nb\") pod \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.594471 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crkwp\" (UniqueName: \"kubernetes.io/projected/3f57f828-9ffd-4caf-872a-fc3fd1913df5-kube-api-access-crkwp\") pod \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.594546 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-dns-svc\") pod \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.594601 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-sb\") pod \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\" (UID: \"3f57f828-9ffd-4caf-872a-fc3fd1913df5\") " Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.615984 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f57f828-9ffd-4caf-872a-fc3fd1913df5-kube-api-access-crkwp" (OuterVolumeSpecName: "kube-api-access-crkwp") pod "3f57f828-9ffd-4caf-872a-fc3fd1913df5" (UID: "3f57f828-9ffd-4caf-872a-fc3fd1913df5"). InnerVolumeSpecName "kube-api-access-crkwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.637452 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zjjvp" event={"ID":"3f57f828-9ffd-4caf-872a-fc3fd1913df5","Type":"ContainerDied","Data":"bf083e2f2ea224805a96fe73d8de4e530cad6e3ec5a2d1698ebb229b1d4eaecc"} Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.637525 4788 scope.go:117] "RemoveContainer" containerID="4ae05de9ee9dfbfd7f28da6127b6ad354efa3958ccfb741c52c9a125d68d2d1d" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.637612 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zjjvp" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.660027 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3f57f828-9ffd-4caf-872a-fc3fd1913df5" (UID: "3f57f828-9ffd-4caf-872a-fc3fd1913df5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.661196 4788 scope.go:117] "RemoveContainer" containerID="b66badae53fe5050595a9fc745c9c05d52a914ddb1db479fee6dbe97f1709580" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.661476 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3f57f828-9ffd-4caf-872a-fc3fd1913df5" (UID: "3f57f828-9ffd-4caf-872a-fc3fd1913df5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.670762 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3f57f828-9ffd-4caf-872a-fc3fd1913df5" (UID: "3f57f828-9ffd-4caf-872a-fc3fd1913df5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.680051 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-config" (OuterVolumeSpecName: "config") pod "3f57f828-9ffd-4caf-872a-fc3fd1913df5" (UID: "3f57f828-9ffd-4caf-872a-fc3fd1913df5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.697567 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.697631 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.697650 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crkwp\" (UniqueName: \"kubernetes.io/projected/3f57f828-9ffd-4caf-872a-fc3fd1913df5-kube-api-access-crkwp\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.697661 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.697672 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3f57f828-9ffd-4caf-872a-fc3fd1913df5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.757403 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-79446"] Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.912518 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c44p7"] Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.922124 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-cmdtx"] Oct 10 15:04:59 crc kubenswrapper[4788]: I1010 15:04:59.938473 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-sclcr"] Oct 10 15:04:59 crc kubenswrapper[4788]: W1010 15:04:59.946283 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5b098f0_5965_4ab3_a641_e301b2f599ed.slice/crio-74373cfe43df49c9464f9b6af0e828c6ec446c1b204f4b98785cba866261dffb WatchSource:0}: Error finding container 74373cfe43df49c9464f9b6af0e828c6ec446c1b204f4b98785cba866261dffb: Status 404 returned error can't find the container with id 74373cfe43df49c9464f9b6af0e828c6ec446c1b204f4b98785cba866261dffb Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.001986 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zjjvp"] Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.009379 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zjjvp"] Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.245834 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" path="/var/lib/kubelet/pods/3f57f828-9ffd-4caf-872a-fc3fd1913df5/volumes" Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.655210 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c44p7" event={"ID":"1063e9ab-7517-4a5e-8865-f7827c52245a","Type":"ContainerStarted","Data":"abbbd83b9330b2ca671f12c78a96221a3b38e4af520c021b0b9fcda847d3d31a"} Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.672384 4788 generic.go:334] "Generic (PLEG): container finished" podID="5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8" containerID="f4727ecaab5a16f71e927909b73eabb95375907d30bc07c7b99154a067738b3b" exitCode=0 Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.672505 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-79446" event={"ID":"5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8","Type":"ContainerDied","Data":"f4727ecaab5a16f71e927909b73eabb95375907d30bc07c7b99154a067738b3b"} Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.672561 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-79446" event={"ID":"5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8","Type":"ContainerStarted","Data":"c72f76e6ab403c97f851505ac6288c8b796098ac0798d9aa5b70cf062cc7ebc9"} Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.674904 4788 generic.go:334] "Generic (PLEG): container finished" podID="679c6797-8be3-49ee-ada2-c2ea883db467" containerID="6d31303eab8e8df640fcd2caa8d708c4dba787f55933092d07939f7904769a25" exitCode=0 Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.674983 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cmdtx" event={"ID":"679c6797-8be3-49ee-ada2-c2ea883db467","Type":"ContainerDied","Data":"6d31303eab8e8df640fcd2caa8d708c4dba787f55933092d07939f7904769a25"} Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.675019 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cmdtx" event={"ID":"679c6797-8be3-49ee-ada2-c2ea883db467","Type":"ContainerStarted","Data":"8b36d97612951851a6d8c51c35ce3796bf8c251de3c9d188c187faf11b236b25"} Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.677050 4788 generic.go:334] "Generic (PLEG): container finished" podID="b5b098f0-5965-4ab3-a641-e301b2f599ed" containerID="4718475c27409de0e7a18cce2d0271c55efe267c0167c7e7a90691925b3c008d" exitCode=0 Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.677109 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sclcr" event={"ID":"b5b098f0-5965-4ab3-a641-e301b2f599ed","Type":"ContainerDied","Data":"4718475c27409de0e7a18cce2d0271c55efe267c0167c7e7a90691925b3c008d"} Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.677133 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sclcr" event={"ID":"b5b098f0-5965-4ab3-a641-e301b2f599ed","Type":"ContainerStarted","Data":"74373cfe43df49c9464f9b6af0e828c6ec446c1b204f4b98785cba866261dffb"} Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.679806 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-cpld9" event={"ID":"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02","Type":"ContainerStarted","Data":"7e850bcecdc4369a4a000f3af74b9e10aa254c1fd8e0b4cd9c9bce7f62cee54e"} Oct 10 15:05:00 crc kubenswrapper[4788]: I1010 15:05:00.724279 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-cpld9" podStartSLOduration=3.413209166 podStartE2EDuration="17.724257513s" podCreationTimestamp="2025-10-10 15:04:43 +0000 UTC" firstStartedPulling="2025-10-10 15:04:44.992660944 +0000 UTC m=+1187.442376512" lastFinishedPulling="2025-10-10 15:04:59.303709311 +0000 UTC m=+1201.753424859" observedRunningTime="2025-10-10 15:05:00.720441 +0000 UTC m=+1203.170156548" watchObservedRunningTime="2025-10-10 15:05:00.724257513 +0000 UTC m=+1203.173973061" Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.722744 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cmdtx" event={"ID":"679c6797-8be3-49ee-ada2-c2ea883db467","Type":"ContainerDied","Data":"8b36d97612951851a6d8c51c35ce3796bf8c251de3c9d188c187faf11b236b25"} Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.723930 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b36d97612951851a6d8c51c35ce3796bf8c251de3c9d188c187faf11b236b25" Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.727673 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-sclcr" event={"ID":"b5b098f0-5965-4ab3-a641-e301b2f599ed","Type":"ContainerDied","Data":"74373cfe43df49c9464f9b6af0e828c6ec446c1b204f4b98785cba866261dffb"} Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.727722 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74373cfe43df49c9464f9b6af0e828c6ec446c1b204f4b98785cba866261dffb" Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.729967 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-79446" event={"ID":"5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8","Type":"ContainerDied","Data":"c72f76e6ab403c97f851505ac6288c8b796098ac0798d9aa5b70cf062cc7ebc9"} Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.730004 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c72f76e6ab403c97f851505ac6288c8b796098ac0798d9aa5b70cf062cc7ebc9" Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.878223 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-79446" Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.906991 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sclcr" Oct 10 15:05:04 crc kubenswrapper[4788]: I1010 15:05:04.928010 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cmdtx" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.028760 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfjx2\" (UniqueName: \"kubernetes.io/projected/b5b098f0-5965-4ab3-a641-e301b2f599ed-kube-api-access-hfjx2\") pod \"b5b098f0-5965-4ab3-a641-e301b2f599ed\" (UID: \"b5b098f0-5965-4ab3-a641-e301b2f599ed\") " Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.028842 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxg72\" (UniqueName: \"kubernetes.io/projected/5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8-kube-api-access-qxg72\") pod \"5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8\" (UID: \"5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8\") " Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.028956 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wdnb\" (UniqueName: \"kubernetes.io/projected/679c6797-8be3-49ee-ada2-c2ea883db467-kube-api-access-4wdnb\") pod \"679c6797-8be3-49ee-ada2-c2ea883db467\" (UID: \"679c6797-8be3-49ee-ada2-c2ea883db467\") " Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.033852 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/679c6797-8be3-49ee-ada2-c2ea883db467-kube-api-access-4wdnb" (OuterVolumeSpecName: "kube-api-access-4wdnb") pod "679c6797-8be3-49ee-ada2-c2ea883db467" (UID: "679c6797-8be3-49ee-ada2-c2ea883db467"). InnerVolumeSpecName "kube-api-access-4wdnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.033883 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8-kube-api-access-qxg72" (OuterVolumeSpecName: "kube-api-access-qxg72") pod "5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8" (UID: "5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8"). InnerVolumeSpecName "kube-api-access-qxg72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.033857 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b098f0-5965-4ab3-a641-e301b2f599ed-kube-api-access-hfjx2" (OuterVolumeSpecName: "kube-api-access-hfjx2") pod "b5b098f0-5965-4ab3-a641-e301b2f599ed" (UID: "b5b098f0-5965-4ab3-a641-e301b2f599ed"). InnerVolumeSpecName "kube-api-access-hfjx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.132323 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfjx2\" (UniqueName: \"kubernetes.io/projected/b5b098f0-5965-4ab3-a641-e301b2f599ed-kube-api-access-hfjx2\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.132369 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxg72\" (UniqueName: \"kubernetes.io/projected/5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8-kube-api-access-qxg72\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.132380 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wdnb\" (UniqueName: \"kubernetes.io/projected/679c6797-8be3-49ee-ada2-c2ea883db467-kube-api-access-4wdnb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.742969 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c44p7" event={"ID":"1063e9ab-7517-4a5e-8865-f7827c52245a","Type":"ContainerStarted","Data":"f2e78e25b19d6852b324ff847adff58351738286f1c416abcc7aab836feedf39"} Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.743045 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-79446" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.743079 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cmdtx" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.743080 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-sclcr" Oct 10 15:05:05 crc kubenswrapper[4788]: I1010 15:05:05.763834 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-c44p7" podStartSLOduration=10.007578872 podStartE2EDuration="14.76380589s" podCreationTimestamp="2025-10-10 15:04:51 +0000 UTC" firstStartedPulling="2025-10-10 15:04:59.926653539 +0000 UTC m=+1202.376369087" lastFinishedPulling="2025-10-10 15:05:04.682880527 +0000 UTC m=+1207.132596105" observedRunningTime="2025-10-10 15:05:05.761667802 +0000 UTC m=+1208.211383360" watchObservedRunningTime="2025-10-10 15:05:05.76380589 +0000 UTC m=+1208.213521448" Oct 10 15:05:06 crc kubenswrapper[4788]: I1010 15:05:06.759732 4788 generic.go:334] "Generic (PLEG): container finished" podID="a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" containerID="7e850bcecdc4369a4a000f3af74b9e10aa254c1fd8e0b4cd9c9bce7f62cee54e" exitCode=0 Oct 10 15:05:06 crc kubenswrapper[4788]: I1010 15:05:06.759786 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-cpld9" event={"ID":"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02","Type":"ContainerDied","Data":"7e850bcecdc4369a4a000f3af74b9e10aa254c1fd8e0b4cd9c9bce7f62cee54e"} Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.253432 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-cpld9" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.395261 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-combined-ca-bundle\") pod \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.395743 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-db-sync-config-data\") pod \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.395804 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whx96\" (UniqueName: \"kubernetes.io/projected/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-kube-api-access-whx96\") pod \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.395910 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-config-data\") pod \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\" (UID: \"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02\") " Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.403793 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-kube-api-access-whx96" (OuterVolumeSpecName: "kube-api-access-whx96") pod "a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" (UID: "a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02"). InnerVolumeSpecName "kube-api-access-whx96". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.403896 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" (UID: "a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.434080 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" (UID: "a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.466806 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-config-data" (OuterVolumeSpecName: "config-data") pod "a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" (UID: "a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.497646 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.497681 4788 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.497692 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whx96\" (UniqueName: \"kubernetes.io/projected/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-kube-api-access-whx96\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.497707 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.784461 4788 generic.go:334] "Generic (PLEG): container finished" podID="1063e9ab-7517-4a5e-8865-f7827c52245a" containerID="f2e78e25b19d6852b324ff847adff58351738286f1c416abcc7aab836feedf39" exitCode=0 Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.784566 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c44p7" event={"ID":"1063e9ab-7517-4a5e-8865-f7827c52245a","Type":"ContainerDied","Data":"f2e78e25b19d6852b324ff847adff58351738286f1c416abcc7aab836feedf39"} Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.786920 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-cpld9" event={"ID":"a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02","Type":"ContainerDied","Data":"62b8035c071cdcbd5b5240c6bf7d7966d3ad283f4c95239ae9094fea421c691e"} Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.786969 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62b8035c071cdcbd5b5240c6bf7d7966d3ad283f4c95239ae9094fea421c691e" Oct 10 15:05:08 crc kubenswrapper[4788]: I1010 15:05:08.786984 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-cpld9" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.242972 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2bq2d"] Oct 10 15:05:09 crc kubenswrapper[4788]: E1010 15:05:09.243386 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="679c6797-8be3-49ee-ada2-c2ea883db467" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243406 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="679c6797-8be3-49ee-ada2-c2ea883db467" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: E1010 15:05:09.243422 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerName="dnsmasq-dns" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243429 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerName="dnsmasq-dns" Oct 10 15:05:09 crc kubenswrapper[4788]: E1010 15:05:09.243444 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b098f0-5965-4ab3-a641-e301b2f599ed" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243449 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b098f0-5965-4ab3-a641-e301b2f599ed" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: E1010 15:05:09.243467 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerName="init" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243476 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerName="init" Oct 10 15:05:09 crc kubenswrapper[4788]: E1010 15:05:09.243486 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243493 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: E1010 15:05:09.243518 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" containerName="glance-db-sync" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243526 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" containerName="glance-db-sync" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243740 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f57f828-9ffd-4caf-872a-fc3fd1913df5" containerName="dnsmasq-dns" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243755 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b098f0-5965-4ab3-a641-e301b2f599ed" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243775 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="679c6797-8be3-49ee-ada2-c2ea883db467" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243786 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" containerName="glance-db-sync" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.243804 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8" containerName="mariadb-database-create" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.244960 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.272368 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2bq2d"] Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.320917 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.322009 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p58g\" (UniqueName: \"kubernetes.io/projected/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-kube-api-access-2p58g\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.322279 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.322625 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.329282 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.329438 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-config\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.430884 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-config\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.430969 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.431018 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p58g\" (UniqueName: \"kubernetes.io/projected/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-kube-api-access-2p58g\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.431071 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.431156 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.431184 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.432643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.432685 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.432699 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-config\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.432713 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.433521 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.453361 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p58g\" (UniqueName: \"kubernetes.io/projected/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-kube-api-access-2p58g\") pod \"dnsmasq-dns-7ff5475cc9-2bq2d\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:09 crc kubenswrapper[4788]: I1010 15:05:09.562023 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:10 crc kubenswrapper[4788]: I1010 15:05:10.056040 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2bq2d"] Oct 10 15:05:10 crc kubenswrapper[4788]: I1010 15:05:10.821630 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" event={"ID":"e8884a03-8fb5-4b08-9ecc-716c8d9f6922","Type":"ContainerStarted","Data":"d331182a96328461a62609b902c4762ce220f80dc74169488977df109ffa3fcc"} Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.017088 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-59ec-account-create-dzhlt"] Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.019368 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-59ec-account-create-dzhlt" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.022323 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.028451 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-59ec-account-create-dzhlt"] Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.166237 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqj4v\" (UniqueName: \"kubernetes.io/projected/df37fb0e-106c-4fe4-8652-cad4d132d205-kube-api-access-sqj4v\") pod \"barbican-59ec-account-create-dzhlt\" (UID: \"df37fb0e-106c-4fe4-8652-cad4d132d205\") " pod="openstack/barbican-59ec-account-create-dzhlt" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.211984 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-b986-account-create-hr255"] Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.213053 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b986-account-create-hr255" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.221482 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.226818 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b986-account-create-hr255"] Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.268115 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pnzl\" (UniqueName: \"kubernetes.io/projected/eae89614-8508-4323-adf1-0692cedda9ca-kube-api-access-2pnzl\") pod \"cinder-b986-account-create-hr255\" (UID: \"eae89614-8508-4323-adf1-0692cedda9ca\") " pod="openstack/cinder-b986-account-create-hr255" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.268336 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqj4v\" (UniqueName: \"kubernetes.io/projected/df37fb0e-106c-4fe4-8652-cad4d132d205-kube-api-access-sqj4v\") pod \"barbican-59ec-account-create-dzhlt\" (UID: \"df37fb0e-106c-4fe4-8652-cad4d132d205\") " pod="openstack/barbican-59ec-account-create-dzhlt" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.289070 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqj4v\" (UniqueName: \"kubernetes.io/projected/df37fb0e-106c-4fe4-8652-cad4d132d205-kube-api-access-sqj4v\") pod \"barbican-59ec-account-create-dzhlt\" (UID: \"df37fb0e-106c-4fe4-8652-cad4d132d205\") " pod="openstack/barbican-59ec-account-create-dzhlt" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.315856 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ab26-account-create-24bhk"] Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.317456 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab26-account-create-24bhk" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.320509 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.338783 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ab26-account-create-24bhk"] Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.340258 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-59ec-account-create-dzhlt" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.369552 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pnzl\" (UniqueName: \"kubernetes.io/projected/eae89614-8508-4323-adf1-0692cedda9ca-kube-api-access-2pnzl\") pod \"cinder-b986-account-create-hr255\" (UID: \"eae89614-8508-4323-adf1-0692cedda9ca\") " pod="openstack/cinder-b986-account-create-hr255" Oct 10 15:05:11 crc kubenswrapper[4788]: I1010 15:05:11.391536 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pnzl\" (UniqueName: \"kubernetes.io/projected/eae89614-8508-4323-adf1-0692cedda9ca-kube-api-access-2pnzl\") pod \"cinder-b986-account-create-hr255\" (UID: \"eae89614-8508-4323-adf1-0692cedda9ca\") " pod="openstack/cinder-b986-account-create-hr255" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:11.471655 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqkmr\" (UniqueName: \"kubernetes.io/projected/e2c43dc7-c35c-4406-9825-ae24803eba5e-kube-api-access-cqkmr\") pod \"neutron-ab26-account-create-24bhk\" (UID: \"e2c43dc7-c35c-4406-9825-ae24803eba5e\") " pod="openstack/neutron-ab26-account-create-24bhk" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:11.535960 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b986-account-create-hr255" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:11.575985 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqkmr\" (UniqueName: \"kubernetes.io/projected/e2c43dc7-c35c-4406-9825-ae24803eba5e-kube-api-access-cqkmr\") pod \"neutron-ab26-account-create-24bhk\" (UID: \"e2c43dc7-c35c-4406-9825-ae24803eba5e\") " pod="openstack/neutron-ab26-account-create-24bhk" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:11.599735 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqkmr\" (UniqueName: \"kubernetes.io/projected/e2c43dc7-c35c-4406-9825-ae24803eba5e-kube-api-access-cqkmr\") pod \"neutron-ab26-account-create-24bhk\" (UID: \"e2c43dc7-c35c-4406-9825-ae24803eba5e\") " pod="openstack/neutron-ab26-account-create-24bhk" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:11.774184 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab26-account-create-24bhk" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.058127 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c44p7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.212842 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmxkq\" (UniqueName: \"kubernetes.io/projected/1063e9ab-7517-4a5e-8865-f7827c52245a-kube-api-access-gmxkq\") pod \"1063e9ab-7517-4a5e-8865-f7827c52245a\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.213425 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-combined-ca-bundle\") pod \"1063e9ab-7517-4a5e-8865-f7827c52245a\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.213585 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-config-data\") pod \"1063e9ab-7517-4a5e-8865-f7827c52245a\" (UID: \"1063e9ab-7517-4a5e-8865-f7827c52245a\") " Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.227445 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1063e9ab-7517-4a5e-8865-f7827c52245a-kube-api-access-gmxkq" (OuterVolumeSpecName: "kube-api-access-gmxkq") pod "1063e9ab-7517-4a5e-8865-f7827c52245a" (UID: "1063e9ab-7517-4a5e-8865-f7827c52245a"). InnerVolumeSpecName "kube-api-access-gmxkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.242828 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1063e9ab-7517-4a5e-8865-f7827c52245a" (UID: "1063e9ab-7517-4a5e-8865-f7827c52245a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.265930 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-config-data" (OuterVolumeSpecName: "config-data") pod "1063e9ab-7517-4a5e-8865-f7827c52245a" (UID: "1063e9ab-7517-4a5e-8865-f7827c52245a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.316747 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.316804 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1063e9ab-7517-4a5e-8865-f7827c52245a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.316823 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmxkq\" (UniqueName: \"kubernetes.io/projected/1063e9ab-7517-4a5e-8865-f7827c52245a-kube-api-access-gmxkq\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.937743 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c44p7" event={"ID":"1063e9ab-7517-4a5e-8865-f7827c52245a","Type":"ContainerDied","Data":"abbbd83b9330b2ca671f12c78a96221a3b38e4af520c021b0b9fcda847d3d31a"} Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.937791 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abbbd83b9330b2ca671f12c78a96221a3b38e4af520c021b0b9fcda847d3d31a" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:13.937868 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c44p7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.370434 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2bq2d"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.389896 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hv2rb"] Oct 10 15:05:15 crc kubenswrapper[4788]: E1010 15:05:14.395188 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1063e9ab-7517-4a5e-8865-f7827c52245a" containerName="keystone-db-sync" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.395245 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1063e9ab-7517-4a5e-8865-f7827c52245a" containerName="keystone-db-sync" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.395633 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1063e9ab-7517-4a5e-8865-f7827c52245a" containerName="keystone-db-sync" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.396700 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.399952 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.400235 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.401084 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wg8bv" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.401255 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.408066 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hv2rb"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.438507 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.440802 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.462184 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543682 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543738 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-config\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543769 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543807 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-scripts\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543833 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-combined-ca-bundle\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543859 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vwl5\" (UniqueName: \"kubernetes.io/projected/1f012621-c857-4266-b866-0aa89c4d7492-kube-api-access-8vwl5\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543918 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543943 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-fernet-keys\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.543966 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.544006 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-config-data\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.544035 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-credential-keys\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.544060 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpxvz\" (UniqueName: \"kubernetes.io/projected/b1ea560d-30dc-4b9c-9ce8-523b3582de38-kube-api-access-zpxvz\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646133 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-config-data\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646196 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-credential-keys\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646222 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpxvz\" (UniqueName: \"kubernetes.io/projected/b1ea560d-30dc-4b9c-9ce8-523b3582de38-kube-api-access-zpxvz\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646278 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646303 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-config\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646324 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646352 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-scripts\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646372 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-combined-ca-bundle\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646392 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vwl5\" (UniqueName: \"kubernetes.io/projected/1f012621-c857-4266-b866-0aa89c4d7492-kube-api-access-8vwl5\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646436 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646456 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-fernet-keys\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.646476 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.647584 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.648031 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.648255 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-config\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.648497 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.648513 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.651115 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-combined-ca-bundle\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.651554 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-fernet-keys\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.654052 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-config-data\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.654075 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-credential-keys\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.654413 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-scripts\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.683631 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpxvz\" (UniqueName: \"kubernetes.io/projected/b1ea560d-30dc-4b9c-9ce8-523b3582de38-kube-api-access-zpxvz\") pod \"keystone-bootstrap-hv2rb\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.686880 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vwl5\" (UniqueName: \"kubernetes.io/projected/1f012621-c857-4266-b866-0aa89c4d7492-kube-api-access-8vwl5\") pod \"dnsmasq-dns-5c5cc7c5ff-l5f4m\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.716094 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.753690 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.754480 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.790020 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9wvx7"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.791587 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.800463 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-cmdqc"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.801623 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.806400 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.806447 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gz79p" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.806629 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.812261 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9wvx7"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.830572 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cmdqc"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.854680 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.854749 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-config\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.854789 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.854830 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.854901 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.854925 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhpxz\" (UniqueName: \"kubernetes.io/projected/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-kube-api-access-nhpxz\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958038 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-logs\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958741 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958780 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-combined-ca-bundle\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958809 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-config\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958840 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958893 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958926 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42slb\" (UniqueName: \"kubernetes.io/projected/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-kube-api-access-42slb\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958954 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-scripts\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.958996 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.959018 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhpxz\" (UniqueName: \"kubernetes.io/projected/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-kube-api-access-nhpxz\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.959037 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-config-data\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.960094 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.961784 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-config\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.961798 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.961990 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.967737 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.971206 4788 generic.go:334] "Generic (PLEG): container finished" podID="e8884a03-8fb5-4b08-9ecc-716c8d9f6922" containerID="25bec1b027307b474077a9ef1998f32d09e31b765bae146bdea7a08afde2a2f9" exitCode=0 Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.971278 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" event={"ID":"e8884a03-8fb5-4b08-9ecc-716c8d9f6922","Type":"ContainerDied","Data":"25bec1b027307b474077a9ef1998f32d09e31b765bae146bdea7a08afde2a2f9"} Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:14.994778 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhpxz\" (UniqueName: \"kubernetes.io/projected/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-kube-api-access-nhpxz\") pod \"dnsmasq-dns-8b5c85b87-9wvx7\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.060877 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-combined-ca-bundle\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.060981 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42slb\" (UniqueName: \"kubernetes.io/projected/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-kube-api-access-42slb\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.061008 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-scripts\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.061068 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-config-data\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.061096 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-logs\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.063228 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-logs\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.065183 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-scripts\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.065857 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-config-data\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.065889 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-combined-ca-bundle\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.079868 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42slb\" (UniqueName: \"kubernetes.io/projected/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-kube-api-access-42slb\") pod \"placement-db-sync-cmdqc\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.178972 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:15 crc kubenswrapper[4788]: E1010 15:05:15.189200 4788 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 10 15:05:15 crc kubenswrapper[4788]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/e8884a03-8fb5-4b08-9ecc-716c8d9f6922/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 10 15:05:15 crc kubenswrapper[4788]: > podSandboxID="d331182a96328461a62609b902c4762ce220f80dc74169488977df109ffa3fcc" Oct 10 15:05:15 crc kubenswrapper[4788]: E1010 15:05:15.189343 4788 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 10 15:05:15 crc kubenswrapper[4788]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64h68fh95h595h67fh597hfch57ch68fh5ffh6hf4h689h659h569h65bh67bh65dh594h64h5d6hd8h5bfh9fh5c4h676h5cdh56h8bh569h664h645q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2p58g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7ff5475cc9-2bq2d_openstack(e8884a03-8fb5-4b08-9ecc-716c8d9f6922): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/e8884a03-8fb5-4b08-9ecc-716c8d9f6922/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 10 15:05:15 crc kubenswrapper[4788]: > logger="UnhandledError" Oct 10 15:05:15 crc kubenswrapper[4788]: E1010 15:05:15.190474 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/e8884a03-8fb5-4b08-9ecc-716c8d9f6922/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" podUID="e8884a03-8fb5-4b08-9ecc-716c8d9f6922" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.237509 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.518457 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.523112 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.530839 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.531458 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.531478 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.531672 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-887mn" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.541552 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.571719 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.571846 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-scripts\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.571950 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-logs\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.572116 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.572244 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k499b\" (UniqueName: \"kubernetes.io/projected/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-kube-api-access-k499b\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.572759 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-config-data\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.572799 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.572867 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.656006 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.658281 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.665544 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.666491 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.673926 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-logs\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.673988 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.674018 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k499b\" (UniqueName: \"kubernetes.io/projected/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-kube-api-access-k499b\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.674077 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-config-data\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.674094 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.674121 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.674184 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.674203 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-scripts\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.674893 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.676731 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.680768 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-scripts\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.683070 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-logs\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.683774 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.686851 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.694653 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.697244 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-config-data\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.708275 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k499b\" (UniqueName: \"kubernetes.io/projected/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-kube-api-access-k499b\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.764128 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.844020 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.884188 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.884269 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.884302 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.884333 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.884352 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.884401 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-logs\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.884417 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.884439 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2zns\" (UniqueName: \"kubernetes.io/projected/5ec40923-13a6-4ff9-8448-691c9afbb375-kube-api-access-x2zns\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991205 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991466 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991502 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991520 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991577 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-logs\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991591 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991610 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2zns\" (UniqueName: \"kubernetes.io/projected/5ec40923-13a6-4ff9-8448-691c9afbb375-kube-api-access-x2zns\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991657 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:15 crc kubenswrapper[4788]: I1010 15:05:15.991824 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.005561 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.016442 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-logs\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.019194 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.021970 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.022114 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.037090 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.055096 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2zns\" (UniqueName: \"kubernetes.io/projected/5ec40923-13a6-4ff9-8448-691c9afbb375-kube-api-access-x2zns\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.069037 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ab26-account-create-24bhk"] Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.108363 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m"] Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.117417 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.147302 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.225246 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b986-account-create-hr255"] Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.286396 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cmdqc"] Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.305790 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hv2rb"] Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.330329 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-59ec-account-create-dzhlt"] Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.343485 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9wvx7"] Oct 10 15:05:16 crc kubenswrapper[4788]: W1010 15:05:16.448718 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf37fb0e_106c_4fe4_8652_cad4d132d205.slice/crio-b1bdce292c1998de149a206bdf8be58aafd721f9b2fbf437df250d498c27cd32 WatchSource:0}: Error finding container b1bdce292c1998de149a206bdf8be58aafd721f9b2fbf437df250d498c27cd32: Status 404 returned error can't find the container with id b1bdce292c1998de149a206bdf8be58aafd721f9b2fbf437df250d498c27cd32 Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.585120 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.707878 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-svc\") pod \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.708387 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-swift-storage-0\") pod \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.708417 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p58g\" (UniqueName: \"kubernetes.io/projected/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-kube-api-access-2p58g\") pod \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.708440 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-nb\") pod \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.708474 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-sb\") pod \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.708535 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-config\") pod \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\" (UID: \"e8884a03-8fb5-4b08-9ecc-716c8d9f6922\") " Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.746825 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-kube-api-access-2p58g" (OuterVolumeSpecName: "kube-api-access-2p58g") pod "e8884a03-8fb5-4b08-9ecc-716c8d9f6922" (UID: "e8884a03-8fb5-4b08-9ecc-716c8d9f6922"). InnerVolumeSpecName "kube-api-access-2p58g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.835809 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p58g\" (UniqueName: \"kubernetes.io/projected/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-kube-api-access-2p58g\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.845376 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.885973 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e8884a03-8fb5-4b08-9ecc-716c8d9f6922" (UID: "e8884a03-8fb5-4b08-9ecc-716c8d9f6922"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.899041 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8884a03-8fb5-4b08-9ecc-716c8d9f6922" (UID: "e8884a03-8fb5-4b08-9ecc-716c8d9f6922"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.899238 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-config" (OuterVolumeSpecName: "config") pod "e8884a03-8fb5-4b08-9ecc-716c8d9f6922" (UID: "e8884a03-8fb5-4b08-9ecc-716c8d9f6922"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.901717 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8884a03-8fb5-4b08-9ecc-716c8d9f6922" (UID: "e8884a03-8fb5-4b08-9ecc-716c8d9f6922"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.920905 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8884a03-8fb5-4b08-9ecc-716c8d9f6922" (UID: "e8884a03-8fb5-4b08-9ecc-716c8d9f6922"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.941374 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.941411 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.941421 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.941451 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:16 crc kubenswrapper[4788]: I1010 15:05:16.941463 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8884a03-8fb5-4b08-9ecc-716c8d9f6922-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.011543 4788 generic.go:334] "Generic (PLEG): container finished" podID="1f012621-c857-4266-b866-0aa89c4d7492" containerID="2178ee74e5d9d905ab673a5408b8ef581be914e732ba592839f26f0d1bf141fe" exitCode=0 Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.011604 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" event={"ID":"1f012621-c857-4266-b866-0aa89c4d7492","Type":"ContainerDied","Data":"2178ee74e5d9d905ab673a5408b8ef581be914e732ba592839f26f0d1bf141fe"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.011665 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" event={"ID":"1f012621-c857-4266-b866-0aa89c4d7492","Type":"ContainerStarted","Data":"1668baedc34e5730125990cd0c211d78457a358eb8e0fe89f26ebab0d34388e1"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.013922 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cmdqc" event={"ID":"d8d7fc66-5e71-4858-96ef-fad2fc9189f9","Type":"ContainerStarted","Data":"ad22b0bf0b4221b66ae24fe425cb5f6ac852e2eee0938e1a8599a497811ed02d"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.025017 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" event={"ID":"e8884a03-8fb5-4b08-9ecc-716c8d9f6922","Type":"ContainerDied","Data":"d331182a96328461a62609b902c4762ce220f80dc74169488977df109ffa3fcc"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.025066 4788 scope.go:117] "RemoveContainer" containerID="25bec1b027307b474077a9ef1998f32d09e31b765bae146bdea7a08afde2a2f9" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.025207 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-2bq2d" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.035412 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hv2rb" event={"ID":"b1ea560d-30dc-4b9c-9ce8-523b3582de38","Type":"ContainerStarted","Data":"59d414f971d9049d275be0eca426c5c85df49ffbd6a3195e6c288e33645927bc"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.035481 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hv2rb" event={"ID":"b1ea560d-30dc-4b9c-9ce8-523b3582de38","Type":"ContainerStarted","Data":"62a57eca2bcb75051832491b980799be237a630a4c3d9128abfbedace7425ae3"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.037096 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cfe60e91-9b3b-454d-bc7a-39bded91bf1b","Type":"ContainerStarted","Data":"7012380be584fdcac1bd3545ef6faf8e2664bb34f155201cf4a920048e0f5214"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.039122 4788 generic.go:334] "Generic (PLEG): container finished" podID="eae89614-8508-4323-adf1-0692cedda9ca" containerID="7924633e2252255a2f8fe0feed662aecd204a9fa8c9987dfa08717028cec687a" exitCode=0 Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.039186 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b986-account-create-hr255" event={"ID":"eae89614-8508-4323-adf1-0692cedda9ca","Type":"ContainerDied","Data":"7924633e2252255a2f8fe0feed662aecd204a9fa8c9987dfa08717028cec687a"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.039204 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b986-account-create-hr255" event={"ID":"eae89614-8508-4323-adf1-0692cedda9ca","Type":"ContainerStarted","Data":"00341bb060ed01473c0735bb2fe598d3883a489d0dfea7b9b6bff23f97f42ef2"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.040573 4788 generic.go:334] "Generic (PLEG): container finished" podID="df37fb0e-106c-4fe4-8652-cad4d132d205" containerID="e71c6da1b980b5831fd3e520526064c0e485bbe80ec60e1dc2ac9a1be89dab0e" exitCode=0 Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.040621 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-59ec-account-create-dzhlt" event={"ID":"df37fb0e-106c-4fe4-8652-cad4d132d205","Type":"ContainerDied","Data":"e71c6da1b980b5831fd3e520526064c0e485bbe80ec60e1dc2ac9a1be89dab0e"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.040637 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-59ec-account-create-dzhlt" event={"ID":"df37fb0e-106c-4fe4-8652-cad4d132d205","Type":"ContainerStarted","Data":"b1bdce292c1998de149a206bdf8be58aafd721f9b2fbf437df250d498c27cd32"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.041409 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" event={"ID":"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3","Type":"ContainerStarted","Data":"c1039dab291f6e117ae135158db3ad708e176c16435b3a138a2cc488a6d37718"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.044157 4788 generic.go:334] "Generic (PLEG): container finished" podID="e2c43dc7-c35c-4406-9825-ae24803eba5e" containerID="8830b296e336ae0bb205059ae2ae08d16b5ce3f0c6cc07992145bafa1eddc0ce" exitCode=0 Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.044189 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab26-account-create-24bhk" event={"ID":"e2c43dc7-c35c-4406-9825-ae24803eba5e","Type":"ContainerDied","Data":"8830b296e336ae0bb205059ae2ae08d16b5ce3f0c6cc07992145bafa1eddc0ce"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.044204 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab26-account-create-24bhk" event={"ID":"e2c43dc7-c35c-4406-9825-ae24803eba5e","Type":"ContainerStarted","Data":"9b476ec07ae33d90b6a5bf23cb9ff6dac7c1022d40c635457f6c5927b3857c5a"} Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.051667 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hv2rb" podStartSLOduration=3.051651247 podStartE2EDuration="3.051651247s" podCreationTimestamp="2025-10-10 15:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:17.048367459 +0000 UTC m=+1219.498083007" watchObservedRunningTime="2025-10-10 15:05:17.051651247 +0000 UTC m=+1219.501366785" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.313706 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2bq2d"] Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.334240 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-2bq2d"] Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.389652 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.498269 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.564012 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.635022 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:05:17 crc kubenswrapper[4788]: E1010 15:05:17.635465 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8884a03-8fb5-4b08-9ecc-716c8d9f6922" containerName="init" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.635479 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8884a03-8fb5-4b08-9ecc-716c8d9f6922" containerName="init" Oct 10 15:05:17 crc kubenswrapper[4788]: E1010 15:05:17.635514 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f012621-c857-4266-b866-0aa89c4d7492" containerName="init" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.635520 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f012621-c857-4266-b866-0aa89c4d7492" containerName="init" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.635698 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8884a03-8fb5-4b08-9ecc-716c8d9f6922" containerName="init" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.635709 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f012621-c857-4266-b866-0aa89c4d7492" containerName="init" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.637433 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.641872 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.642067 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.662775 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vwl5\" (UniqueName: \"kubernetes.io/projected/1f012621-c857-4266-b866-0aa89c4d7492-kube-api-access-8vwl5\") pod \"1f012621-c857-4266-b866-0aa89c4d7492\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.662955 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-swift-storage-0\") pod \"1f012621-c857-4266-b866-0aa89c4d7492\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.663111 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-config\") pod \"1f012621-c857-4266-b866-0aa89c4d7492\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.663193 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-svc\") pod \"1f012621-c857-4266-b866-0aa89c4d7492\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.663250 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-sb\") pod \"1f012621-c857-4266-b866-0aa89c4d7492\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.663294 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-nb\") pod \"1f012621-c857-4266-b866-0aa89c4d7492\" (UID: \"1f012621-c857-4266-b866-0aa89c4d7492\") " Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.664551 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.676476 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f012621-c857-4266-b866-0aa89c4d7492-kube-api-access-8vwl5" (OuterVolumeSpecName: "kube-api-access-8vwl5") pod "1f012621-c857-4266-b866-0aa89c4d7492" (UID: "1f012621-c857-4266-b866-0aa89c4d7492"). InnerVolumeSpecName "kube-api-access-8vwl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.724354 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1f012621-c857-4266-b866-0aa89c4d7492" (UID: "1f012621-c857-4266-b866-0aa89c4d7492"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.747713 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.754993 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1f012621-c857-4266-b866-0aa89c4d7492" (UID: "1f012621-c857-4266-b866-0aa89c4d7492"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.765910 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.765980 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-run-httpd\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.766008 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klvsv\" (UniqueName: \"kubernetes.io/projected/82934fb8-b2e0-4d17-b071-e06e747515f4-kube-api-access-klvsv\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.766079 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.766119 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-log-httpd\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.766218 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-scripts\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.766263 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-config-data\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.766328 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.766343 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vwl5\" (UniqueName: \"kubernetes.io/projected/1f012621-c857-4266-b866-0aa89c4d7492-kube-api-access-8vwl5\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.766356 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.794019 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1f012621-c857-4266-b866-0aa89c4d7492" (UID: "1f012621-c857-4266-b866-0aa89c4d7492"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.796034 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.829125 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1f012621-c857-4266-b866-0aa89c4d7492" (UID: "1f012621-c857-4266-b866-0aa89c4d7492"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.830924 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-config" (OuterVolumeSpecName: "config") pod "1f012621-c857-4266-b866-0aa89c4d7492" (UID: "1f012621-c857-4266-b866-0aa89c4d7492"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868108 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-scripts\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868355 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-config-data\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868433 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868506 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-run-httpd\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868558 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klvsv\" (UniqueName: \"kubernetes.io/projected/82934fb8-b2e0-4d17-b071-e06e747515f4-kube-api-access-klvsv\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868654 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868693 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-log-httpd\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868827 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868848 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.868908 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f012621-c857-4266-b866-0aa89c4d7492-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.873347 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-log-httpd\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.873418 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-run-httpd\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.881239 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-scripts\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.881710 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.887992 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.888194 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-config-data\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.897879 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klvsv\" (UniqueName: \"kubernetes.io/projected/82934fb8-b2e0-4d17-b071-e06e747515f4-kube-api-access-klvsv\") pod \"ceilometer-0\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " pod="openstack/ceilometer-0" Oct 10 15:05:17 crc kubenswrapper[4788]: I1010 15:05:17.956024 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.066974 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cfe60e91-9b3b-454d-bc7a-39bded91bf1b","Type":"ContainerStarted","Data":"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780"} Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.075334 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.077000 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m" event={"ID":"1f012621-c857-4266-b866-0aa89c4d7492","Type":"ContainerDied","Data":"1668baedc34e5730125990cd0c211d78457a358eb8e0fe89f26ebab0d34388e1"} Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.077065 4788 scope.go:117] "RemoveContainer" containerID="2178ee74e5d9d905ab673a5408b8ef581be914e732ba592839f26f0d1bf141fe" Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.089649 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ec40923-13a6-4ff9-8448-691c9afbb375","Type":"ContainerStarted","Data":"c356a3952e0724ee94c785476e8334e56a0df81c332631c6af54784b5c3a3f83"} Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.093652 4788 generic.go:334] "Generic (PLEG): container finished" podID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" containerID="bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e" exitCode=0 Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.093702 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" event={"ID":"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3","Type":"ContainerDied","Data":"bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e"} Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.198962 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m"] Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.496826 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8884a03-8fb5-4b08-9ecc-716c8d9f6922" path="/var/lib/kubelet/pods/e8884a03-8fb5-4b08-9ecc-716c8d9f6922/volumes" Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.498390 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-l5f4m"] Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.929229 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b986-account-create-hr255" Oct 10 15:05:18 crc kubenswrapper[4788]: I1010 15:05:18.960982 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:05:18 crc kubenswrapper[4788]: W1010 15:05:18.971980 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82934fb8_b2e0_4d17_b071_e06e747515f4.slice/crio-ae09df5e4acbc2dd0cff3489cdc76943478bfc338ab99b35c526e1ccdf22edd5 WatchSource:0}: Error finding container ae09df5e4acbc2dd0cff3489cdc76943478bfc338ab99b35c526e1ccdf22edd5: Status 404 returned error can't find the container with id ae09df5e4acbc2dd0cff3489cdc76943478bfc338ab99b35c526e1ccdf22edd5 Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.023419 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-59ec-account-create-dzhlt" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.038559 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pnzl\" (UniqueName: \"kubernetes.io/projected/eae89614-8508-4323-adf1-0692cedda9ca-kube-api-access-2pnzl\") pod \"eae89614-8508-4323-adf1-0692cedda9ca\" (UID: \"eae89614-8508-4323-adf1-0692cedda9ca\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.072237 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae89614-8508-4323-adf1-0692cedda9ca-kube-api-access-2pnzl" (OuterVolumeSpecName: "kube-api-access-2pnzl") pod "eae89614-8508-4323-adf1-0692cedda9ca" (UID: "eae89614-8508-4323-adf1-0692cedda9ca"). InnerVolumeSpecName "kube-api-access-2pnzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.074621 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab26-account-create-24bhk" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.148829 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqj4v\" (UniqueName: \"kubernetes.io/projected/df37fb0e-106c-4fe4-8652-cad4d132d205-kube-api-access-sqj4v\") pod \"df37fb0e-106c-4fe4-8652-cad4d132d205\" (UID: \"df37fb0e-106c-4fe4-8652-cad4d132d205\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.149322 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqkmr\" (UniqueName: \"kubernetes.io/projected/e2c43dc7-c35c-4406-9825-ae24803eba5e-kube-api-access-cqkmr\") pod \"e2c43dc7-c35c-4406-9825-ae24803eba5e\" (UID: \"e2c43dc7-c35c-4406-9825-ae24803eba5e\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.150049 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pnzl\" (UniqueName: \"kubernetes.io/projected/eae89614-8508-4323-adf1-0692cedda9ca-kube-api-access-2pnzl\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.150349 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" event={"ID":"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3","Type":"ContainerStarted","Data":"42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9"} Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.158347 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.163021 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df37fb0e-106c-4fe4-8652-cad4d132d205-kube-api-access-sqj4v" (OuterVolumeSpecName: "kube-api-access-sqj4v") pod "df37fb0e-106c-4fe4-8652-cad4d132d205" (UID: "df37fb0e-106c-4fe4-8652-cad4d132d205"). InnerVolumeSpecName "kube-api-access-sqj4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.163128 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2c43dc7-c35c-4406-9825-ae24803eba5e-kube-api-access-cqkmr" (OuterVolumeSpecName: "kube-api-access-cqkmr") pod "e2c43dc7-c35c-4406-9825-ae24803eba5e" (UID: "e2c43dc7-c35c-4406-9825-ae24803eba5e"). InnerVolumeSpecName "kube-api-access-cqkmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.168337 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ab26-account-create-24bhk" event={"ID":"e2c43dc7-c35c-4406-9825-ae24803eba5e","Type":"ContainerDied","Data":"9b476ec07ae33d90b6a5bf23cb9ff6dac7c1022d40c635457f6c5927b3857c5a"} Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.168406 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b476ec07ae33d90b6a5bf23cb9ff6dac7c1022d40c635457f6c5927b3857c5a" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.168486 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ab26-account-create-24bhk" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.180750 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerName="glance-log" containerID="cri-o://3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780" gracePeriod=30 Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.180867 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerName="glance-httpd" containerID="cri-o://9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821" gracePeriod=30 Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.187553 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b986-account-create-hr255" event={"ID":"eae89614-8508-4323-adf1-0692cedda9ca","Type":"ContainerDied","Data":"00341bb060ed01473c0735bb2fe598d3883a489d0dfea7b9b6bff23f97f42ef2"} Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.187633 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00341bb060ed01473c0735bb2fe598d3883a489d0dfea7b9b6bff23f97f42ef2" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.187732 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b986-account-create-hr255" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.194195 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82934fb8-b2e0-4d17-b071-e06e747515f4","Type":"ContainerStarted","Data":"ae09df5e4acbc2dd0cff3489cdc76943478bfc338ab99b35c526e1ccdf22edd5"} Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.198782 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-59ec-account-create-dzhlt" event={"ID":"df37fb0e-106c-4fe4-8652-cad4d132d205","Type":"ContainerDied","Data":"b1bdce292c1998de149a206bdf8be58aafd721f9b2fbf437df250d498c27cd32"} Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.198840 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1bdce292c1998de149a206bdf8be58aafd721f9b2fbf437df250d498c27cd32" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.198914 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-59ec-account-create-dzhlt" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.217625 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" podStartSLOduration=5.217600543 podStartE2EDuration="5.217600543s" podCreationTimestamp="2025-10-10 15:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:19.176493914 +0000 UTC m=+1221.626209462" watchObservedRunningTime="2025-10-10 15:05:19.217600543 +0000 UTC m=+1221.667316111" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.217831 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.217802709 podStartE2EDuration="5.217802709s" podCreationTimestamp="2025-10-10 15:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:19.208592101 +0000 UTC m=+1221.658307649" watchObservedRunningTime="2025-10-10 15:05:19.217802709 +0000 UTC m=+1221.667518267" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.251706 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqj4v\" (UniqueName: \"kubernetes.io/projected/df37fb0e-106c-4fe4-8652-cad4d132d205-kube-api-access-sqj4v\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.251747 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqkmr\" (UniqueName: \"kubernetes.io/projected/e2c43dc7-c35c-4406-9825-ae24803eba5e-kube-api-access-cqkmr\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.831981 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.883164 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-scripts\") pod \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.883269 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-httpd-run\") pod \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.883355 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-logs\") pod \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.883382 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k499b\" (UniqueName: \"kubernetes.io/projected/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-kube-api-access-k499b\") pod \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.883490 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-config-data\") pod \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.883554 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-combined-ca-bundle\") pod \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.883625 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-public-tls-certs\") pod \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.883692 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\" (UID: \"cfe60e91-9b3b-454d-bc7a-39bded91bf1b\") " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.885664 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cfe60e91-9b3b-454d-bc7a-39bded91bf1b" (UID: "cfe60e91-9b3b-454d-bc7a-39bded91bf1b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.885683 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-logs" (OuterVolumeSpecName: "logs") pod "cfe60e91-9b3b-454d-bc7a-39bded91bf1b" (UID: "cfe60e91-9b3b-454d-bc7a-39bded91bf1b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.893620 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "cfe60e91-9b3b-454d-bc7a-39bded91bf1b" (UID: "cfe60e91-9b3b-454d-bc7a-39bded91bf1b"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.895862 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-scripts" (OuterVolumeSpecName: "scripts") pod "cfe60e91-9b3b-454d-bc7a-39bded91bf1b" (UID: "cfe60e91-9b3b-454d-bc7a-39bded91bf1b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.902615 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-kube-api-access-k499b" (OuterVolumeSpecName: "kube-api-access-k499b") pod "cfe60e91-9b3b-454d-bc7a-39bded91bf1b" (UID: "cfe60e91-9b3b-454d-bc7a-39bded91bf1b"). InnerVolumeSpecName "kube-api-access-k499b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.960043 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfe60e91-9b3b-454d-bc7a-39bded91bf1b" (UID: "cfe60e91-9b3b-454d-bc7a-39bded91bf1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.968782 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cfe60e91-9b3b-454d-bc7a-39bded91bf1b" (UID: "cfe60e91-9b3b-454d-bc7a-39bded91bf1b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.985905 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.986078 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.986176 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.986265 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.986326 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.986376 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:19 crc kubenswrapper[4788]: I1010 15:05:19.986425 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k499b\" (UniqueName: \"kubernetes.io/projected/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-kube-api-access-k499b\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.016073 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.050438 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-config-data" (OuterVolumeSpecName: "config-data") pod "cfe60e91-9b3b-454d-bc7a-39bded91bf1b" (UID: "cfe60e91-9b3b-454d-bc7a-39bded91bf1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.089069 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfe60e91-9b3b-454d-bc7a-39bded91bf1b-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.089117 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.211692 4788 generic.go:334] "Generic (PLEG): container finished" podID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerID="9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821" exitCode=143 Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.211748 4788 generic.go:334] "Generic (PLEG): container finished" podID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerID="3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780" exitCode=143 Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.211814 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cfe60e91-9b3b-454d-bc7a-39bded91bf1b","Type":"ContainerDied","Data":"9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821"} Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.211847 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cfe60e91-9b3b-454d-bc7a-39bded91bf1b","Type":"ContainerDied","Data":"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780"} Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.211857 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cfe60e91-9b3b-454d-bc7a-39bded91bf1b","Type":"ContainerDied","Data":"7012380be584fdcac1bd3545ef6faf8e2664bb34f155201cf4a920048e0f5214"} Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.211889 4788 scope.go:117] "RemoveContainer" containerID="9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.212045 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.222677 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerName="glance-log" containerID="cri-o://46526274ab62d4fd12d9f773f797314b3009761087682b5b23fabd0b2bbbd9f6" gracePeriod=30 Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.222929 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ec40923-13a6-4ff9-8448-691c9afbb375","Type":"ContainerStarted","Data":"cfd7c6459b60fe350dcb282312ed8eebdbf32a11a8aae45ab3f22975d95bdd8d"} Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.222954 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ec40923-13a6-4ff9-8448-691c9afbb375","Type":"ContainerStarted","Data":"46526274ab62d4fd12d9f773f797314b3009761087682b5b23fabd0b2bbbd9f6"} Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.223209 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerName="glance-httpd" containerID="cri-o://cfd7c6459b60fe350dcb282312ed8eebdbf32a11a8aae45ab3f22975d95bdd8d" gracePeriod=30 Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.250099 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.250083858 podStartE2EDuration="6.250083858s" podCreationTimestamp="2025-10-10 15:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:20.245652558 +0000 UTC m=+1222.695368106" watchObservedRunningTime="2025-10-10 15:05:20.250083858 +0000 UTC m=+1222.699799406" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.263702 4788 scope.go:117] "RemoveContainer" containerID="3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.334083 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f012621-c857-4266-b866-0aa89c4d7492" path="/var/lib/kubelet/pods/1f012621-c857-4266-b866-0aa89c4d7492/volumes" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.364521 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.366918 4788 scope.go:117] "RemoveContainer" containerID="9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821" Oct 10 15:05:20 crc kubenswrapper[4788]: E1010 15:05:20.374967 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821\": container with ID starting with 9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821 not found: ID does not exist" containerID="9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.375011 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821"} err="failed to get container status \"9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821\": rpc error: code = NotFound desc = could not find container \"9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821\": container with ID starting with 9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821 not found: ID does not exist" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.375055 4788 scope.go:117] "RemoveContainer" containerID="3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780" Oct 10 15:05:20 crc kubenswrapper[4788]: E1010 15:05:20.403777 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780\": container with ID starting with 3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780 not found: ID does not exist" containerID="3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.403828 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780"} err="failed to get container status \"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780\": rpc error: code = NotFound desc = could not find container \"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780\": container with ID starting with 3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780 not found: ID does not exist" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.403853 4788 scope.go:117] "RemoveContainer" containerID="9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.411971 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821"} err="failed to get container status \"9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821\": rpc error: code = NotFound desc = could not find container \"9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821\": container with ID starting with 9a8406676c827c76aba51da894c418bd07c1974bb5c6653b4ca02a1e971ca821 not found: ID does not exist" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.412018 4788 scope.go:117] "RemoveContainer" containerID="3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.412929 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780"} err="failed to get container status \"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780\": rpc error: code = NotFound desc = could not find container \"3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780\": container with ID starting with 3c2783ac7cd3016f9e57acd8879d58a1cd99aee5b998f1eaa62ac2e31b8d4780 not found: ID does not exist" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.417117 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.430741 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:20 crc kubenswrapper[4788]: E1010 15:05:20.431280 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerName="glance-httpd" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431302 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerName="glance-httpd" Oct 10 15:05:20 crc kubenswrapper[4788]: E1010 15:05:20.431333 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2c43dc7-c35c-4406-9825-ae24803eba5e" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431340 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2c43dc7-c35c-4406-9825-ae24803eba5e" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: E1010 15:05:20.431352 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerName="glance-log" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431357 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerName="glance-log" Oct 10 15:05:20 crc kubenswrapper[4788]: E1010 15:05:20.431373 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae89614-8508-4323-adf1-0692cedda9ca" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431380 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae89614-8508-4323-adf1-0692cedda9ca" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: E1010 15:05:20.431393 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df37fb0e-106c-4fe4-8652-cad4d132d205" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431402 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="df37fb0e-106c-4fe4-8652-cad4d132d205" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431568 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="df37fb0e-106c-4fe4-8652-cad4d132d205" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431578 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2c43dc7-c35c-4406-9825-ae24803eba5e" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431586 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerName="glance-log" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431605 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae89614-8508-4323-adf1-0692cedda9ca" containerName="mariadb-account-create" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.431616 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" containerName="glance-httpd" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.432641 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.438608 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.438836 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.443804 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.517076 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.517462 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.517610 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-logs\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.517744 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vdcm\" (UniqueName: \"kubernetes.io/projected/dc0199c4-f6d9-44e6-a78e-e651c09d0447-kube-api-access-6vdcm\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.517768 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.517864 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.517921 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.518023 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.620410 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.620466 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.620505 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.620523 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-logs\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.620573 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vdcm\" (UniqueName: \"kubernetes.io/projected/dc0199c4-f6d9-44e6-a78e-e651c09d0447-kube-api-access-6vdcm\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.620588 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.620625 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.620648 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.621047 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.622196 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-logs\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.623108 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.626428 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.627162 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.627384 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-config-data\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.627612 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-scripts\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.672808 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vdcm\" (UniqueName: \"kubernetes.io/projected/dc0199c4-f6d9-44e6-a78e-e651c09d0447-kube-api-access-6vdcm\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.708625 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " pod="openstack/glance-default-external-api-0" Oct 10 15:05:20 crc kubenswrapper[4788]: I1010 15:05:20.768811 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.317909 4788 generic.go:334] "Generic (PLEG): container finished" podID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerID="cfd7c6459b60fe350dcb282312ed8eebdbf32a11a8aae45ab3f22975d95bdd8d" exitCode=0 Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.318215 4788 generic.go:334] "Generic (PLEG): container finished" podID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerID="46526274ab62d4fd12d9f773f797314b3009761087682b5b23fabd0b2bbbd9f6" exitCode=143 Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.318049 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ec40923-13a6-4ff9-8448-691c9afbb375","Type":"ContainerDied","Data":"cfd7c6459b60fe350dcb282312ed8eebdbf32a11a8aae45ab3f22975d95bdd8d"} Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.318327 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ec40923-13a6-4ff9-8448-691c9afbb375","Type":"ContainerDied","Data":"46526274ab62d4fd12d9f773f797314b3009761087682b5b23fabd0b2bbbd9f6"} Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.371876 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-sbnwc"] Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.373049 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.376629 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9ghf5" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.376863 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.402985 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sbnwc"] Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.441425 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-combined-ca-bundle\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.441477 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-db-sync-config-data\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.441574 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvfrj\" (UniqueName: \"kubernetes.io/projected/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-kube-api-access-hvfrj\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.477970 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.543609 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-combined-ca-bundle\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.543668 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-db-sync-config-data\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.543744 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvfrj\" (UniqueName: \"kubernetes.io/projected/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-kube-api-access-hvfrj\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.554671 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-combined-ca-bundle\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.558597 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-db-sync-config-data\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.561652 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvfrj\" (UniqueName: \"kubernetes.io/projected/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-kube-api-access-hvfrj\") pod \"barbican-db-sync-sbnwc\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.641284 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-cnk7r"] Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.642977 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.646358 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.646556 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.646659 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2fdbx" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.654036 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cnk7r"] Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.704095 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.751186 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-db-sync-config-data\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.751393 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-config-data\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.751436 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7wm4\" (UniqueName: \"kubernetes.io/projected/63cf2515-0dd5-4f1f-b70e-08a095284b53-kube-api-access-p7wm4\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.751563 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-scripts\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.751639 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63cf2515-0dd5-4f1f-b70e-08a095284b53-etc-machine-id\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.751727 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-combined-ca-bundle\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.753488 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-mxj9t"] Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.763974 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.765668 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mxj9t"] Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.767700 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.768051 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.768241 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4q8dr" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.853943 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63cf2515-0dd5-4f1f-b70e-08a095284b53-etc-machine-id\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854044 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-combined-ca-bundle\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854091 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngkjh\" (UniqueName: \"kubernetes.io/projected/c95690b5-a444-46f4-98b6-997531cad4e8-kube-api-access-ngkjh\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854118 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-combined-ca-bundle\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854193 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-db-sync-config-data\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854227 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-config\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854264 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-config-data\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854288 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7wm4\" (UniqueName: \"kubernetes.io/projected/63cf2515-0dd5-4f1f-b70e-08a095284b53-kube-api-access-p7wm4\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854318 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-scripts\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.854790 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63cf2515-0dd5-4f1f-b70e-08a095284b53-etc-machine-id\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.861550 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-db-sync-config-data\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.862212 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-combined-ca-bundle\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.866619 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-config-data\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.873353 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-scripts\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.882844 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7wm4\" (UniqueName: \"kubernetes.io/projected/63cf2515-0dd5-4f1f-b70e-08a095284b53-kube-api-access-p7wm4\") pod \"cinder-db-sync-cnk7r\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.955862 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngkjh\" (UniqueName: \"kubernetes.io/projected/c95690b5-a444-46f4-98b6-997531cad4e8-kube-api-access-ngkjh\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.956787 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-combined-ca-bundle\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.956981 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-config\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.960811 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-combined-ca-bundle\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.964710 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-config\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:21 crc kubenswrapper[4788]: I1010 15:05:21.973956 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngkjh\" (UniqueName: \"kubernetes.io/projected/c95690b5-a444-46f4-98b6-997531cad4e8-kube-api-access-ngkjh\") pod \"neutron-db-sync-mxj9t\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:22 crc kubenswrapper[4788]: I1010 15:05:22.028805 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:05:22 crc kubenswrapper[4788]: I1010 15:05:22.119989 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:22 crc kubenswrapper[4788]: I1010 15:05:22.252210 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe60e91-9b3b-454d-bc7a-39bded91bf1b" path="/var/lib/kubelet/pods/cfe60e91-9b3b-454d-bc7a-39bded91bf1b/volumes" Oct 10 15:05:22 crc kubenswrapper[4788]: I1010 15:05:22.343594 4788 generic.go:334] "Generic (PLEG): container finished" podID="b1ea560d-30dc-4b9c-9ce8-523b3582de38" containerID="59d414f971d9049d275be0eca426c5c85df49ffbd6a3195e6c288e33645927bc" exitCode=0 Oct 10 15:05:22 crc kubenswrapper[4788]: I1010 15:05:22.343639 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hv2rb" event={"ID":"b1ea560d-30dc-4b9c-9ce8-523b3582de38","Type":"ContainerDied","Data":"59d414f971d9049d275be0eca426c5c85df49ffbd6a3195e6c288e33645927bc"} Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.504954 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.585038 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-httpd-run\") pod \"5ec40923-13a6-4ff9-8448-691c9afbb375\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.585582 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-combined-ca-bundle\") pod \"5ec40923-13a6-4ff9-8448-691c9afbb375\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.585557 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5ec40923-13a6-4ff9-8448-691c9afbb375" (UID: "5ec40923-13a6-4ff9-8448-691c9afbb375"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.586019 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-logs" (OuterVolumeSpecName: "logs") pod "5ec40923-13a6-4ff9-8448-691c9afbb375" (UID: "5ec40923-13a6-4ff9-8448-691c9afbb375"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.585652 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-logs\") pod \"5ec40923-13a6-4ff9-8448-691c9afbb375\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.586901 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-internal-tls-certs\") pod \"5ec40923-13a6-4ff9-8448-691c9afbb375\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.586983 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-config-data\") pod \"5ec40923-13a6-4ff9-8448-691c9afbb375\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.587021 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2zns\" (UniqueName: \"kubernetes.io/projected/5ec40923-13a6-4ff9-8448-691c9afbb375-kube-api-access-x2zns\") pod \"5ec40923-13a6-4ff9-8448-691c9afbb375\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.587072 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"5ec40923-13a6-4ff9-8448-691c9afbb375\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.587198 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-scripts\") pod \"5ec40923-13a6-4ff9-8448-691c9afbb375\" (UID: \"5ec40923-13a6-4ff9-8448-691c9afbb375\") " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.587845 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.587862 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ec40923-13a6-4ff9-8448-691c9afbb375-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.593464 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-scripts" (OuterVolumeSpecName: "scripts") pod "5ec40923-13a6-4ff9-8448-691c9afbb375" (UID: "5ec40923-13a6-4ff9-8448-691c9afbb375"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.593906 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "5ec40923-13a6-4ff9-8448-691c9afbb375" (UID: "5ec40923-13a6-4ff9-8448-691c9afbb375"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.605547 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec40923-13a6-4ff9-8448-691c9afbb375-kube-api-access-x2zns" (OuterVolumeSpecName: "kube-api-access-x2zns") pod "5ec40923-13a6-4ff9-8448-691c9afbb375" (UID: "5ec40923-13a6-4ff9-8448-691c9afbb375"). InnerVolumeSpecName "kube-api-access-x2zns". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.628872 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ec40923-13a6-4ff9-8448-691c9afbb375" (UID: "5ec40923-13a6-4ff9-8448-691c9afbb375"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.650351 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-config-data" (OuterVolumeSpecName: "config-data") pod "5ec40923-13a6-4ff9-8448-691c9afbb375" (UID: "5ec40923-13a6-4ff9-8448-691c9afbb375"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.672810 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5ec40923-13a6-4ff9-8448-691c9afbb375" (UID: "5ec40923-13a6-4ff9-8448-691c9afbb375"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.693159 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.693208 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.693219 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.693229 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2zns\" (UniqueName: \"kubernetes.io/projected/5ec40923-13a6-4ff9-8448-691c9afbb375-kube-api-access-x2zns\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.693279 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.693290 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec40923-13a6-4ff9-8448-691c9afbb375-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.717655 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 10 15:05:23 crc kubenswrapper[4788]: I1010 15:05:23.795226 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.363352 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc0199c4-f6d9-44e6-a78e-e651c09d0447","Type":"ContainerStarted","Data":"cac065e6108d17f60762aed680df0db20232700bb1d2a11f245629d395933d2a"} Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.365636 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5ec40923-13a6-4ff9-8448-691c9afbb375","Type":"ContainerDied","Data":"c356a3952e0724ee94c785476e8334e56a0df81c332631c6af54784b5c3a3f83"} Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.365669 4788 scope.go:117] "RemoveContainer" containerID="cfd7c6459b60fe350dcb282312ed8eebdbf32a11a8aae45ab3f22975d95bdd8d" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.365731 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.394628 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.412338 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.428759 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:24 crc kubenswrapper[4788]: E1010 15:05:24.429446 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerName="glance-log" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.429475 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerName="glance-log" Oct 10 15:05:24 crc kubenswrapper[4788]: E1010 15:05:24.429514 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerName="glance-httpd" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.429524 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerName="glance-httpd" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.429766 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerName="glance-log" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.429795 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" containerName="glance-httpd" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.431287 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.434171 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.434367 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.454941 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.510999 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.511056 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-logs\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.511098 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.511119 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jff6\" (UniqueName: \"kubernetes.io/projected/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-kube-api-access-7jff6\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.511362 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.511419 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.511489 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.511585 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.622777 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.622836 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.622881 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.622941 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.623008 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.623051 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-logs\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.623105 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.623130 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jff6\" (UniqueName: \"kubernetes.io/projected/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-kube-api-access-7jff6\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.623899 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.624441 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-logs\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.624756 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.627513 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.628328 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.628955 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.632729 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.654334 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jff6\" (UniqueName: \"kubernetes.io/projected/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-kube-api-access-7jff6\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.672511 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:05:24 crc kubenswrapper[4788]: I1010 15:05:24.748935 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:25 crc kubenswrapper[4788]: I1010 15:05:25.181374 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:05:25 crc kubenswrapper[4788]: I1010 15:05:25.267599 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-4dn9v"] Oct 10 15:05:25 crc kubenswrapper[4788]: I1010 15:05:25.268094 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" podUID="f654d8a6-43f1-46e7-b046-d1f52af77f48" containerName="dnsmasq-dns" containerID="cri-o://d603d0ab14e8d86d9580feb13b10db4d1d774aab6e488203ffbfdc111c16b2e8" gracePeriod=10 Oct 10 15:05:25 crc kubenswrapper[4788]: I1010 15:05:25.985485 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.065063 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpxvz\" (UniqueName: \"kubernetes.io/projected/b1ea560d-30dc-4b9c-9ce8-523b3582de38-kube-api-access-zpxvz\") pod \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.065214 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-combined-ca-bundle\") pod \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.065270 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-credential-keys\") pod \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.065345 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-scripts\") pod \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.065377 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-fernet-keys\") pod \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.065539 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-config-data\") pod \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\" (UID: \"b1ea560d-30dc-4b9c-9ce8-523b3582de38\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.071249 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1ea560d-30dc-4b9c-9ce8-523b3582de38-kube-api-access-zpxvz" (OuterVolumeSpecName: "kube-api-access-zpxvz") pod "b1ea560d-30dc-4b9c-9ce8-523b3582de38" (UID: "b1ea560d-30dc-4b9c-9ce8-523b3582de38"). InnerVolumeSpecName "kube-api-access-zpxvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.072875 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b1ea560d-30dc-4b9c-9ce8-523b3582de38" (UID: "b1ea560d-30dc-4b9c-9ce8-523b3582de38"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.073757 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-scripts" (OuterVolumeSpecName: "scripts") pod "b1ea560d-30dc-4b9c-9ce8-523b3582de38" (UID: "b1ea560d-30dc-4b9c-9ce8-523b3582de38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.073930 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b1ea560d-30dc-4b9c-9ce8-523b3582de38" (UID: "b1ea560d-30dc-4b9c-9ce8-523b3582de38"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.114307 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1ea560d-30dc-4b9c-9ce8-523b3582de38" (UID: "b1ea560d-30dc-4b9c-9ce8-523b3582de38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.117523 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-config-data" (OuterVolumeSpecName: "config-data") pod "b1ea560d-30dc-4b9c-9ce8-523b3582de38" (UID: "b1ea560d-30dc-4b9c-9ce8-523b3582de38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.167904 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.168292 4788 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.168304 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.168314 4788 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.168325 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ea560d-30dc-4b9c-9ce8-523b3582de38-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.168336 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpxvz\" (UniqueName: \"kubernetes.io/projected/b1ea560d-30dc-4b9c-9ce8-523b3582de38-kube-api-access-zpxvz\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.260923 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ec40923-13a6-4ff9-8448-691c9afbb375" path="/var/lib/kubelet/pods/5ec40923-13a6-4ff9-8448-691c9afbb375/volumes" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.391030 4788 generic.go:334] "Generic (PLEG): container finished" podID="f654d8a6-43f1-46e7-b046-d1f52af77f48" containerID="d603d0ab14e8d86d9580feb13b10db4d1d774aab6e488203ffbfdc111c16b2e8" exitCode=0 Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.391114 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" event={"ID":"f654d8a6-43f1-46e7-b046-d1f52af77f48","Type":"ContainerDied","Data":"d603d0ab14e8d86d9580feb13b10db4d1d774aab6e488203ffbfdc111c16b2e8"} Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.393216 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hv2rb" event={"ID":"b1ea560d-30dc-4b9c-9ce8-523b3582de38","Type":"ContainerDied","Data":"62a57eca2bcb75051832491b980799be237a630a4c3d9128abfbedace7425ae3"} Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.393247 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62a57eca2bcb75051832491b980799be237a630a4c3d9128abfbedace7425ae3" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.393312 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hv2rb" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.473267 4788 scope.go:117] "RemoveContainer" containerID="46526274ab62d4fd12d9f773f797314b3009761087682b5b23fabd0b2bbbd9f6" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.503720 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.677088 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-config\") pod \"f654d8a6-43f1-46e7-b046-d1f52af77f48\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.677279 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-swift-storage-0\") pod \"f654d8a6-43f1-46e7-b046-d1f52af77f48\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.677307 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-nb\") pod \"f654d8a6-43f1-46e7-b046-d1f52af77f48\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.677347 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-svc\") pod \"f654d8a6-43f1-46e7-b046-d1f52af77f48\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.677378 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-sb\") pod \"f654d8a6-43f1-46e7-b046-d1f52af77f48\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.677463 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stcnb\" (UniqueName: \"kubernetes.io/projected/f654d8a6-43f1-46e7-b046-d1f52af77f48-kube-api-access-stcnb\") pod \"f654d8a6-43f1-46e7-b046-d1f52af77f48\" (UID: \"f654d8a6-43f1-46e7-b046-d1f52af77f48\") " Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.684852 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f654d8a6-43f1-46e7-b046-d1f52af77f48-kube-api-access-stcnb" (OuterVolumeSpecName: "kube-api-access-stcnb") pod "f654d8a6-43f1-46e7-b046-d1f52af77f48" (UID: "f654d8a6-43f1-46e7-b046-d1f52af77f48"). InnerVolumeSpecName "kube-api-access-stcnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.780513 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stcnb\" (UniqueName: \"kubernetes.io/projected/f654d8a6-43f1-46e7-b046-d1f52af77f48-kube-api-access-stcnb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.793179 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-config" (OuterVolumeSpecName: "config") pod "f654d8a6-43f1-46e7-b046-d1f52af77f48" (UID: "f654d8a6-43f1-46e7-b046-d1f52af77f48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.873845 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f654d8a6-43f1-46e7-b046-d1f52af77f48" (UID: "f654d8a6-43f1-46e7-b046-d1f52af77f48"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.874001 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f654d8a6-43f1-46e7-b046-d1f52af77f48" (UID: "f654d8a6-43f1-46e7-b046-d1f52af77f48"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.896065 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.896102 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.896112 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.928878 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f654d8a6-43f1-46e7-b046-d1f52af77f48" (UID: "f654d8a6-43f1-46e7-b046-d1f52af77f48"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:26 crc kubenswrapper[4788]: I1010 15:05:26.934909 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f654d8a6-43f1-46e7-b046-d1f52af77f48" (UID: "f654d8a6-43f1-46e7-b046-d1f52af77f48"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:26.997327 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:26.997363 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f654d8a6-43f1-46e7-b046-d1f52af77f48-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.134217 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hv2rb"] Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.155301 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hv2rb"] Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.240298 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-t5zxq"] Oct 10 15:05:27 crc kubenswrapper[4788]: E1010 15:05:27.240711 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f654d8a6-43f1-46e7-b046-d1f52af77f48" containerName="init" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.240725 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f654d8a6-43f1-46e7-b046-d1f52af77f48" containerName="init" Oct 10 15:05:27 crc kubenswrapper[4788]: E1010 15:05:27.240749 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f654d8a6-43f1-46e7-b046-d1f52af77f48" containerName="dnsmasq-dns" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.240755 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f654d8a6-43f1-46e7-b046-d1f52af77f48" containerName="dnsmasq-dns" Oct 10 15:05:27 crc kubenswrapper[4788]: E1010 15:05:27.240780 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ea560d-30dc-4b9c-9ce8-523b3582de38" containerName="keystone-bootstrap" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.240788 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ea560d-30dc-4b9c-9ce8-523b3582de38" containerName="keystone-bootstrap" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.240966 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f654d8a6-43f1-46e7-b046-d1f52af77f48" containerName="dnsmasq-dns" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.240982 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1ea560d-30dc-4b9c-9ce8-523b3582de38" containerName="keystone-bootstrap" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.241611 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.244641 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.244823 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.245584 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wg8bv" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.246057 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.247495 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t5zxq"] Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.302266 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-config-data\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.302334 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwlrn\" (UniqueName: \"kubernetes.io/projected/1586c9ff-8c4a-4946-b524-cf90d3d81db9-kube-api-access-rwlrn\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.302503 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-credential-keys\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.302535 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-scripts\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.302559 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-combined-ca-bundle\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.302598 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-fernet-keys\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.367508 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cnk7r"] Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.405442 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwlrn\" (UniqueName: \"kubernetes.io/projected/1586c9ff-8c4a-4946-b524-cf90d3d81db9-kube-api-access-rwlrn\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.405540 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-credential-keys\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.405564 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-scripts\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.405587 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-combined-ca-bundle\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.405607 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-fernet-keys\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.405667 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-config-data\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.412424 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-scripts\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.413101 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-fernet-keys\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.413801 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-combined-ca-bundle\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.413902 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cnk7r" event={"ID":"63cf2515-0dd5-4f1f-b70e-08a095284b53","Type":"ContainerStarted","Data":"c9f56d84d8e9eded0c1fa3db8d05aa26bcf64626e2c71f61c6ea757de96fd379"} Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.414672 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-config-data\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.416020 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" event={"ID":"f654d8a6-43f1-46e7-b046-d1f52af77f48","Type":"ContainerDied","Data":"2aa2049e9dc7d9707bc9a774f2d300b98e84c8005f4cf85d741f36df4be51240"} Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.416051 4788 scope.go:117] "RemoveContainer" containerID="d603d0ab14e8d86d9580feb13b10db4d1d774aab6e488203ffbfdc111c16b2e8" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.416248 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-4dn9v" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.417336 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-credential-keys\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.430600 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwlrn\" (UniqueName: \"kubernetes.io/projected/1586c9ff-8c4a-4946-b524-cf90d3d81db9-kube-api-access-rwlrn\") pod \"keystone-bootstrap-t5zxq\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.522722 4788 scope.go:117] "RemoveContainer" containerID="4a6299b42981fdad04090c27687cca82ce0e81bf2af962149819d6c845b78f51" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.549186 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-4dn9v"] Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.564478 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-4dn9v"] Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.572799 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.745702 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mxj9t"] Oct 10 15:05:27 crc kubenswrapper[4788]: I1010 15:05:27.769112 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sbnwc"] Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.012790 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:05:28 crc kubenswrapper[4788]: W1010 15:05:28.033851 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9384e19a_7c77_49ca_9dc8_7be525cb4e6c.slice/crio-53bbea773567a0f8e1a99ce19cfca345b382643391fb5abf5ff0dbaff2902598 WatchSource:0}: Error finding container 53bbea773567a0f8e1a99ce19cfca345b382643391fb5abf5ff0dbaff2902598: Status 404 returned error can't find the container with id 53bbea773567a0f8e1a99ce19cfca345b382643391fb5abf5ff0dbaff2902598 Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.127681 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-t5zxq"] Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.256422 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1ea560d-30dc-4b9c-9ce8-523b3582de38" path="/var/lib/kubelet/pods/b1ea560d-30dc-4b9c-9ce8-523b3582de38/volumes" Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.257173 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f654d8a6-43f1-46e7-b046-d1f52af77f48" path="/var/lib/kubelet/pods/f654d8a6-43f1-46e7-b046-d1f52af77f48/volumes" Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.460702 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82934fb8-b2e0-4d17-b071-e06e747515f4","Type":"ContainerStarted","Data":"cd905b3981551c017fee3cce5d78db27acb3a006b0a289d22d1980255e28acb8"} Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.464966 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc0199c4-f6d9-44e6-a78e-e651c09d0447","Type":"ContainerStarted","Data":"e09f1669e3a187300831d30e49a102971bd1cc7640e3796649e5f7fdfbec27e0"} Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.466668 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sbnwc" event={"ID":"1f6f73b5-459a-4490-bac4-426d8b3f2b9e","Type":"ContainerStarted","Data":"fd48b60312edf8cab58c4402b0cc1b46ede9e593913db08ca4abb692eb4bd9e8"} Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.470427 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cmdqc" event={"ID":"d8d7fc66-5e71-4858-96ef-fad2fc9189f9","Type":"ContainerStarted","Data":"907c6c53800cba4ebb91526f1fea5664b224ece7220cf782500a49298e929afc"} Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.490394 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9384e19a-7c77-49ca-9dc8-7be525cb4e6c","Type":"ContainerStarted","Data":"53bbea773567a0f8e1a99ce19cfca345b382643391fb5abf5ff0dbaff2902598"} Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.496379 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-cmdqc" podStartSLOduration=4.257093161 podStartE2EDuration="14.49635578s" podCreationTimestamp="2025-10-10 15:05:14 +0000 UTC" firstStartedPulling="2025-10-10 15:05:16.336829259 +0000 UTC m=+1218.786544807" lastFinishedPulling="2025-10-10 15:05:26.576091868 +0000 UTC m=+1229.025807426" observedRunningTime="2025-10-10 15:05:28.491598392 +0000 UTC m=+1230.941313960" watchObservedRunningTime="2025-10-10 15:05:28.49635578 +0000 UTC m=+1230.946071328" Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.500778 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mxj9t" event={"ID":"c95690b5-a444-46f4-98b6-997531cad4e8","Type":"ContainerStarted","Data":"095bf53a298c6f70b38e43a9a5978cec79e29c28c3436fdae08686599935c768"} Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.500824 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mxj9t" event={"ID":"c95690b5-a444-46f4-98b6-997531cad4e8","Type":"ContainerStarted","Data":"a38d9d266a19cf7f4a04c99abcf2937d97a5d97cf213b7a5abe0505f380dfa9c"} Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.504159 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t5zxq" event={"ID":"1586c9ff-8c4a-4946-b524-cf90d3d81db9","Type":"ContainerStarted","Data":"0cfc94d2a27ea72d41b030e8032cafa5ac984771e805ecaa15994e05b90783d0"} Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.521600 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-mxj9t" podStartSLOduration=7.521581791 podStartE2EDuration="7.521581791s" podCreationTimestamp="2025-10-10 15:05:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:28.518535509 +0000 UTC m=+1230.968251077" watchObservedRunningTime="2025-10-10 15:05:28.521581791 +0000 UTC m=+1230.971297339" Oct 10 15:05:28 crc kubenswrapper[4788]: I1010 15:05:28.551330 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-t5zxq" podStartSLOduration=1.551303033 podStartE2EDuration="1.551303033s" podCreationTimestamp="2025-10-10 15:05:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:28.544595702 +0000 UTC m=+1230.994311250" watchObservedRunningTime="2025-10-10 15:05:28.551303033 +0000 UTC m=+1231.001018581" Oct 10 15:05:29 crc kubenswrapper[4788]: I1010 15:05:29.407537 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:05:29 crc kubenswrapper[4788]: I1010 15:05:29.408350 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:05:29 crc kubenswrapper[4788]: I1010 15:05:29.542793 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t5zxq" event={"ID":"1586c9ff-8c4a-4946-b524-cf90d3d81db9","Type":"ContainerStarted","Data":"bfc0913bcd1ab1498eb49921a29998fe79941d93b829b0a8fb7368a4fa4e2fdd"} Oct 10 15:05:29 crc kubenswrapper[4788]: I1010 15:05:29.559785 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc0199c4-f6d9-44e6-a78e-e651c09d0447","Type":"ContainerStarted","Data":"183b4d966d60fe3ff4c7e99782bb6afa87a35b173ac80cac610a4890b558dd1c"} Oct 10 15:05:29 crc kubenswrapper[4788]: I1010 15:05:29.567248 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9384e19a-7c77-49ca-9dc8-7be525cb4e6c","Type":"ContainerStarted","Data":"48c30f8dd20ca989162fe7a6889da2090b8f3f2925b51f181010597cfe4ab535"} Oct 10 15:05:29 crc kubenswrapper[4788]: I1010 15:05:29.592706 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=9.592687501 podStartE2EDuration="9.592687501s" podCreationTimestamp="2025-10-10 15:05:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:29.587559232 +0000 UTC m=+1232.037274770" watchObservedRunningTime="2025-10-10 15:05:29.592687501 +0000 UTC m=+1232.042403049" Oct 10 15:05:30 crc kubenswrapper[4788]: I1010 15:05:30.589606 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82934fb8-b2e0-4d17-b071-e06e747515f4","Type":"ContainerStarted","Data":"7ecafbd5dc24dad827ac28f9bc73bf2383bcffe7b45d5665ac712e4299074544"} Oct 10 15:05:30 crc kubenswrapper[4788]: I1010 15:05:30.769769 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 15:05:30 crc kubenswrapper[4788]: I1010 15:05:30.769863 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 15:05:30 crc kubenswrapper[4788]: I1010 15:05:30.819063 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 15:05:30 crc kubenswrapper[4788]: I1010 15:05:30.821377 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 15:05:31 crc kubenswrapper[4788]: I1010 15:05:31.608192 4788 generic.go:334] "Generic (PLEG): container finished" podID="d8d7fc66-5e71-4858-96ef-fad2fc9189f9" containerID="907c6c53800cba4ebb91526f1fea5664b224ece7220cf782500a49298e929afc" exitCode=0 Oct 10 15:05:31 crc kubenswrapper[4788]: I1010 15:05:31.608667 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cmdqc" event={"ID":"d8d7fc66-5e71-4858-96ef-fad2fc9189f9","Type":"ContainerDied","Data":"907c6c53800cba4ebb91526f1fea5664b224ece7220cf782500a49298e929afc"} Oct 10 15:05:31 crc kubenswrapper[4788]: I1010 15:05:31.615053 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9384e19a-7c77-49ca-9dc8-7be525cb4e6c","Type":"ContainerStarted","Data":"26dc1f755c12a846444239eb5acf3a8b889a93f2aaa3ed202d6dab1c802f9dae"} Oct 10 15:05:31 crc kubenswrapper[4788]: I1010 15:05:31.615094 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 15:05:31 crc kubenswrapper[4788]: I1010 15:05:31.615160 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 15:05:33 crc kubenswrapper[4788]: I1010 15:05:33.643891 4788 generic.go:334] "Generic (PLEG): container finished" podID="1586c9ff-8c4a-4946-b524-cf90d3d81db9" containerID="bfc0913bcd1ab1498eb49921a29998fe79941d93b829b0a8fb7368a4fa4e2fdd" exitCode=0 Oct 10 15:05:33 crc kubenswrapper[4788]: I1010 15:05:33.644020 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t5zxq" event={"ID":"1586c9ff-8c4a-4946-b524-cf90d3d81db9","Type":"ContainerDied","Data":"bfc0913bcd1ab1498eb49921a29998fe79941d93b829b0a8fb7368a4fa4e2fdd"} Oct 10 15:05:33 crc kubenswrapper[4788]: I1010 15:05:33.669298 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.669273402 podStartE2EDuration="9.669273402s" podCreationTimestamp="2025-10-10 15:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:31.655390375 +0000 UTC m=+1234.105105923" watchObservedRunningTime="2025-10-10 15:05:33.669273402 +0000 UTC m=+1236.118988950" Oct 10 15:05:34 crc kubenswrapper[4788]: I1010 15:05:34.620192 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 15:05:34 crc kubenswrapper[4788]: I1010 15:05:34.633609 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 15:05:34 crc kubenswrapper[4788]: I1010 15:05:34.750798 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:34 crc kubenswrapper[4788]: I1010 15:05:34.750847 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:34 crc kubenswrapper[4788]: I1010 15:05:34.802403 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:34 crc kubenswrapper[4788]: I1010 15:05:34.839393 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:34 crc kubenswrapper[4788]: I1010 15:05:34.878821 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.002311 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42slb\" (UniqueName: \"kubernetes.io/projected/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-kube-api-access-42slb\") pod \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.002441 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-scripts\") pod \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.002498 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-config-data\") pod \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.002767 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-combined-ca-bundle\") pod \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.002808 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-logs\") pod \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\" (UID: \"d8d7fc66-5e71-4858-96ef-fad2fc9189f9\") " Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.003973 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-logs" (OuterVolumeSpecName: "logs") pod "d8d7fc66-5e71-4858-96ef-fad2fc9189f9" (UID: "d8d7fc66-5e71-4858-96ef-fad2fc9189f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.012826 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-scripts" (OuterVolumeSpecName: "scripts") pod "d8d7fc66-5e71-4858-96ef-fad2fc9189f9" (UID: "d8d7fc66-5e71-4858-96ef-fad2fc9189f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.016570 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-kube-api-access-42slb" (OuterVolumeSpecName: "kube-api-access-42slb") pod "d8d7fc66-5e71-4858-96ef-fad2fc9189f9" (UID: "d8d7fc66-5e71-4858-96ef-fad2fc9189f9"). InnerVolumeSpecName "kube-api-access-42slb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.068111 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-config-data" (OuterVolumeSpecName: "config-data") pod "d8d7fc66-5e71-4858-96ef-fad2fc9189f9" (UID: "d8d7fc66-5e71-4858-96ef-fad2fc9189f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.081316 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8d7fc66-5e71-4858-96ef-fad2fc9189f9" (UID: "d8d7fc66-5e71-4858-96ef-fad2fc9189f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.105524 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.105561 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.105576 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.105585 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.105596 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42slb\" (UniqueName: \"kubernetes.io/projected/d8d7fc66-5e71-4858-96ef-fad2fc9189f9-kube-api-access-42slb\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.672585 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cmdqc" event={"ID":"d8d7fc66-5e71-4858-96ef-fad2fc9189f9","Type":"ContainerDied","Data":"ad22b0bf0b4221b66ae24fe425cb5f6ac852e2eee0938e1a8599a497811ed02d"} Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.673006 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad22b0bf0b4221b66ae24fe425cb5f6ac852e2eee0938e1a8599a497811ed02d" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.673116 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cmdqc" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.674014 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.674049 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.994652 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5d565c7b56-z7s74"] Oct 10 15:05:35 crc kubenswrapper[4788]: E1010 15:05:35.995087 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d7fc66-5e71-4858-96ef-fad2fc9189f9" containerName="placement-db-sync" Oct 10 15:05:35 crc kubenswrapper[4788]: I1010 15:05:35.995103 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d7fc66-5e71-4858-96ef-fad2fc9189f9" containerName="placement-db-sync" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:35.995809 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d7fc66-5e71-4858-96ef-fad2fc9189f9" containerName="placement-db-sync" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.048089 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5d565c7b56-z7s74"] Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.048254 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.053054 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.053335 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.053520 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.053562 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gz79p" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.054280 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.136379 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-combined-ca-bundle\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.136444 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-public-tls-certs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.136464 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-internal-tls-certs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.136714 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwbj2\" (UniqueName: \"kubernetes.io/projected/67000f4b-b955-44d7-ad37-3d552d8efd88-kube-api-access-vwbj2\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.136926 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-scripts\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.137328 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-config-data\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.137403 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67000f4b-b955-44d7-ad37-3d552d8efd88-logs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.242387 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-config-data\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.242453 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67000f4b-b955-44d7-ad37-3d552d8efd88-logs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.242494 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-combined-ca-bundle\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.242522 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-public-tls-certs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.242538 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-internal-tls-certs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.242619 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwbj2\" (UniqueName: \"kubernetes.io/projected/67000f4b-b955-44d7-ad37-3d552d8efd88-kube-api-access-vwbj2\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.242698 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-scripts\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.243990 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67000f4b-b955-44d7-ad37-3d552d8efd88-logs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.252090 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-config-data\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.263542 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-internal-tls-certs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.263818 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-scripts\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.264343 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-combined-ca-bundle\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.268798 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwbj2\" (UniqueName: \"kubernetes.io/projected/67000f4b-b955-44d7-ad37-3d552d8efd88-kube-api-access-vwbj2\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.271944 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-public-tls-certs\") pod \"placement-5d565c7b56-z7s74\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.327112 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.382918 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.445590 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-combined-ca-bundle\") pod \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.445792 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-scripts\") pod \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.445944 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-fernet-keys\") pod \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.445981 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwlrn\" (UniqueName: \"kubernetes.io/projected/1586c9ff-8c4a-4946-b524-cf90d3d81db9-kube-api-access-rwlrn\") pod \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.446002 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-config-data\") pod \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.446039 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-credential-keys\") pod \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\" (UID: \"1586c9ff-8c4a-4946-b524-cf90d3d81db9\") " Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.454264 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1586c9ff-8c4a-4946-b524-cf90d3d81db9" (UID: "1586c9ff-8c4a-4946-b524-cf90d3d81db9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.466645 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-scripts" (OuterVolumeSpecName: "scripts") pod "1586c9ff-8c4a-4946-b524-cf90d3d81db9" (UID: "1586c9ff-8c4a-4946-b524-cf90d3d81db9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.466738 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1586c9ff-8c4a-4946-b524-cf90d3d81db9-kube-api-access-rwlrn" (OuterVolumeSpecName: "kube-api-access-rwlrn") pod "1586c9ff-8c4a-4946-b524-cf90d3d81db9" (UID: "1586c9ff-8c4a-4946-b524-cf90d3d81db9"). InnerVolumeSpecName "kube-api-access-rwlrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.468289 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1586c9ff-8c4a-4946-b524-cf90d3d81db9" (UID: "1586c9ff-8c4a-4946-b524-cf90d3d81db9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.477163 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1586c9ff-8c4a-4946-b524-cf90d3d81db9" (UID: "1586c9ff-8c4a-4946-b524-cf90d3d81db9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.501673 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-config-data" (OuterVolumeSpecName: "config-data") pod "1586c9ff-8c4a-4946-b524-cf90d3d81db9" (UID: "1586c9ff-8c4a-4946-b524-cf90d3d81db9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.548391 4788 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.548436 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwlrn\" (UniqueName: \"kubernetes.io/projected/1586c9ff-8c4a-4946-b524-cf90d3d81db9-kube-api-access-rwlrn\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.548447 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.548456 4788 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.548465 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.548472 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1586c9ff-8c4a-4946-b524-cf90d3d81db9-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.689695 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-t5zxq" event={"ID":"1586c9ff-8c4a-4946-b524-cf90d3d81db9","Type":"ContainerDied","Data":"0cfc94d2a27ea72d41b030e8032cafa5ac984771e805ecaa15994e05b90783d0"} Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.689753 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cfc94d2a27ea72d41b030e8032cafa5ac984771e805ecaa15994e05b90783d0" Oct 10 15:05:36 crc kubenswrapper[4788]: I1010 15:05:36.689759 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-t5zxq" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.452621 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7f68d7b48c-24qxb"] Oct 10 15:05:37 crc kubenswrapper[4788]: E1010 15:05:37.453192 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1586c9ff-8c4a-4946-b524-cf90d3d81db9" containerName="keystone-bootstrap" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.453213 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1586c9ff-8c4a-4946-b524-cf90d3d81db9" containerName="keystone-bootstrap" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.453498 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1586c9ff-8c4a-4946-b524-cf90d3d81db9" containerName="keystone-bootstrap" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.454367 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.465712 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.465930 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.465895 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.466041 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wg8bv" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.467161 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.467540 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.477506 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7f68d7b48c-24qxb"] Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.579172 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-scripts\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.579231 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-internal-tls-certs\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.579261 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-credential-keys\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.579335 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrsln\" (UniqueName: \"kubernetes.io/projected/970e7515-bd36-45d9-befb-a0597cc72cbb-kube-api-access-mrsln\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.579405 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-fernet-keys\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.579421 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-public-tls-certs\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.581282 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-config-data\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.581346 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-combined-ca-bundle\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.686517 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-scripts\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.686596 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-internal-tls-certs\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.686628 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-credential-keys\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.686678 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrsln\" (UniqueName: \"kubernetes.io/projected/970e7515-bd36-45d9-befb-a0597cc72cbb-kube-api-access-mrsln\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.686731 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-fernet-keys\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.686757 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-public-tls-certs\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.686791 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-config-data\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.686821 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-combined-ca-bundle\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.701989 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-credential-keys\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.705902 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-public-tls-certs\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.706437 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-config-data\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.710589 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-scripts\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.711025 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-combined-ca-bundle\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.711162 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrsln\" (UniqueName: \"kubernetes.io/projected/970e7515-bd36-45d9-befb-a0597cc72cbb-kube-api-access-mrsln\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.712160 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-internal-tls-certs\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.723031 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-fernet-keys\") pod \"keystone-7f68d7b48c-24qxb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.723954 4788 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.723978 4788 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 15:05:37 crc kubenswrapper[4788]: I1010 15:05:37.802261 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:38 crc kubenswrapper[4788]: I1010 15:05:38.148112 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:38 crc kubenswrapper[4788]: I1010 15:05:38.175680 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 15:05:48 crc kubenswrapper[4788]: E1010 15:05:48.289547 4788 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Oct 10 15:05:48 crc kubenswrapper[4788]: E1010 15:05:48.290012 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-klvsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(82934fb8-b2e0-4d17-b071-e06e747515f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 15:05:49 crc kubenswrapper[4788]: E1010 15:05:49.364939 4788 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 10 15:05:49 crc kubenswrapper[4788]: E1010 15:05:49.365833 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p7wm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-cnk7r_openstack(63cf2515-0dd5-4f1f-b70e-08a095284b53): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 10 15:05:49 crc kubenswrapper[4788]: E1010 15:05:49.367196 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-cnk7r" podUID="63cf2515-0dd5-4f1f-b70e-08a095284b53" Oct 10 15:05:49 crc kubenswrapper[4788]: I1010 15:05:49.687795 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7f68d7b48c-24qxb"] Oct 10 15:05:49 crc kubenswrapper[4788]: W1010 15:05:49.745740 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod970e7515_bd36_45d9_befb_a0597cc72cbb.slice/crio-10d0e0419d7726432bece8b517cfe4d81600b75496c1e64ab430542a46bbcb5b WatchSource:0}: Error finding container 10d0e0419d7726432bece8b517cfe4d81600b75496c1e64ab430542a46bbcb5b: Status 404 returned error can't find the container with id 10d0e0419d7726432bece8b517cfe4d81600b75496c1e64ab430542a46bbcb5b Oct 10 15:05:49 crc kubenswrapper[4788]: I1010 15:05:49.822428 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5d565c7b56-z7s74"] Oct 10 15:05:49 crc kubenswrapper[4788]: W1010 15:05:49.825989 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67000f4b_b955_44d7_ad37_3d552d8efd88.slice/crio-4548a22d7e0d810257f4a467ffef7507c6f03b2a25499966819f1ef5c4552182 WatchSource:0}: Error finding container 4548a22d7e0d810257f4a467ffef7507c6f03b2a25499966819f1ef5c4552182: Status 404 returned error can't find the container with id 4548a22d7e0d810257f4a467ffef7507c6f03b2a25499966819f1ef5c4552182 Oct 10 15:05:49 crc kubenswrapper[4788]: I1010 15:05:49.856767 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sbnwc" event={"ID":"1f6f73b5-459a-4490-bac4-426d8b3f2b9e","Type":"ContainerStarted","Data":"27deb31db11e825d80667118e5461f9ff9e69e6d1632a0d32d3510499e4941b1"} Oct 10 15:05:49 crc kubenswrapper[4788]: I1010 15:05:49.858461 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f68d7b48c-24qxb" event={"ID":"970e7515-bd36-45d9-befb-a0597cc72cbb","Type":"ContainerStarted","Data":"10d0e0419d7726432bece8b517cfe4d81600b75496c1e64ab430542a46bbcb5b"} Oct 10 15:05:49 crc kubenswrapper[4788]: I1010 15:05:49.859640 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d565c7b56-z7s74" event={"ID":"67000f4b-b955-44d7-ad37-3d552d8efd88","Type":"ContainerStarted","Data":"4548a22d7e0d810257f4a467ffef7507c6f03b2a25499966819f1ef5c4552182"} Oct 10 15:05:49 crc kubenswrapper[4788]: I1010 15:05:49.861974 4788 generic.go:334] "Generic (PLEG): container finished" podID="c95690b5-a444-46f4-98b6-997531cad4e8" containerID="095bf53a298c6f70b38e43a9a5978cec79e29c28c3436fdae08686599935c768" exitCode=0 Oct 10 15:05:49 crc kubenswrapper[4788]: I1010 15:05:49.862020 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mxj9t" event={"ID":"c95690b5-a444-46f4-98b6-997531cad4e8","Type":"ContainerDied","Data":"095bf53a298c6f70b38e43a9a5978cec79e29c28c3436fdae08686599935c768"} Oct 10 15:05:49 crc kubenswrapper[4788]: E1010 15:05:49.863720 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-cnk7r" podUID="63cf2515-0dd5-4f1f-b70e-08a095284b53" Oct 10 15:05:49 crc kubenswrapper[4788]: I1010 15:05:49.884737 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-sbnwc" podStartSLOduration=7.322334889 podStartE2EDuration="28.884709476s" podCreationTimestamp="2025-10-10 15:05:21 +0000 UTC" firstStartedPulling="2025-10-10 15:05:27.797942184 +0000 UTC m=+1230.247657732" lastFinishedPulling="2025-10-10 15:05:49.360316761 +0000 UTC m=+1251.810032319" observedRunningTime="2025-10-10 15:05:49.875450675 +0000 UTC m=+1252.325166223" watchObservedRunningTime="2025-10-10 15:05:49.884709476 +0000 UTC m=+1252.334425024" Oct 10 15:05:50 crc kubenswrapper[4788]: I1010 15:05:50.872256 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d565c7b56-z7s74" event={"ID":"67000f4b-b955-44d7-ad37-3d552d8efd88","Type":"ContainerStarted","Data":"e39944b79b0c075079e9a288c1893a4cdbd681cc744af6cff568e6a8933d1c22"} Oct 10 15:05:50 crc kubenswrapper[4788]: I1010 15:05:50.872719 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d565c7b56-z7s74" event={"ID":"67000f4b-b955-44d7-ad37-3d552d8efd88","Type":"ContainerStarted","Data":"f98704522449389d2af1c513f520e6f82ebedfbe293c1adf7038d2875138c32a"} Oct 10 15:05:50 crc kubenswrapper[4788]: I1010 15:05:50.872735 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:50 crc kubenswrapper[4788]: I1010 15:05:50.872745 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:05:50 crc kubenswrapper[4788]: I1010 15:05:50.874994 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f68d7b48c-24qxb" event={"ID":"970e7515-bd36-45d9-befb-a0597cc72cbb","Type":"ContainerStarted","Data":"ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115"} Oct 10 15:05:50 crc kubenswrapper[4788]: I1010 15:05:50.877942 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:05:50 crc kubenswrapper[4788]: I1010 15:05:50.946878 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5d565c7b56-z7s74" podStartSLOduration=15.946854384 podStartE2EDuration="15.946854384s" podCreationTimestamp="2025-10-10 15:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:50.919127096 +0000 UTC m=+1253.368842644" watchObservedRunningTime="2025-10-10 15:05:50.946854384 +0000 UTC m=+1253.396569932" Oct 10 15:05:50 crc kubenswrapper[4788]: I1010 15:05:50.953504 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7f68d7b48c-24qxb" podStartSLOduration=13.953487213 podStartE2EDuration="13.953487213s" podCreationTimestamp="2025-10-10 15:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:05:50.941008866 +0000 UTC m=+1253.390724424" watchObservedRunningTime="2025-10-10 15:05:50.953487213 +0000 UTC m=+1253.403202751" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.405321 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.559530 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngkjh\" (UniqueName: \"kubernetes.io/projected/c95690b5-a444-46f4-98b6-997531cad4e8-kube-api-access-ngkjh\") pod \"c95690b5-a444-46f4-98b6-997531cad4e8\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.559793 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-config\") pod \"c95690b5-a444-46f4-98b6-997531cad4e8\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.559879 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-combined-ca-bundle\") pod \"c95690b5-a444-46f4-98b6-997531cad4e8\" (UID: \"c95690b5-a444-46f4-98b6-997531cad4e8\") " Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.567417 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c95690b5-a444-46f4-98b6-997531cad4e8-kube-api-access-ngkjh" (OuterVolumeSpecName: "kube-api-access-ngkjh") pod "c95690b5-a444-46f4-98b6-997531cad4e8" (UID: "c95690b5-a444-46f4-98b6-997531cad4e8"). InnerVolumeSpecName "kube-api-access-ngkjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.588324 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-config" (OuterVolumeSpecName: "config") pod "c95690b5-a444-46f4-98b6-997531cad4e8" (UID: "c95690b5-a444-46f4-98b6-997531cad4e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.597582 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c95690b5-a444-46f4-98b6-997531cad4e8" (UID: "c95690b5-a444-46f4-98b6-997531cad4e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.663475 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngkjh\" (UniqueName: \"kubernetes.io/projected/c95690b5-a444-46f4-98b6-997531cad4e8-kube-api-access-ngkjh\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.663524 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.663539 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c95690b5-a444-46f4-98b6-997531cad4e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.888350 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mxj9t" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.890467 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mxj9t" event={"ID":"c95690b5-a444-46f4-98b6-997531cad4e8","Type":"ContainerDied","Data":"a38d9d266a19cf7f4a04c99abcf2937d97a5d97cf213b7a5abe0505f380dfa9c"} Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.890619 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a38d9d266a19cf7f4a04c99abcf2937d97a5d97cf213b7a5abe0505f380dfa9c" Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.905392 4788 generic.go:334] "Generic (PLEG): container finished" podID="1f6f73b5-459a-4490-bac4-426d8b3f2b9e" containerID="27deb31db11e825d80667118e5461f9ff9e69e6d1632a0d32d3510499e4941b1" exitCode=0 Oct 10 15:05:51 crc kubenswrapper[4788]: I1010 15:05:51.905675 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sbnwc" event={"ID":"1f6f73b5-459a-4490-bac4-426d8b3f2b9e","Type":"ContainerDied","Data":"27deb31db11e825d80667118e5461f9ff9e69e6d1632a0d32d3510499e4941b1"} Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.068981 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-khsc8"] Oct 10 15:05:52 crc kubenswrapper[4788]: E1010 15:05:52.069482 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95690b5-a444-46f4-98b6-997531cad4e8" containerName="neutron-db-sync" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.069502 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95690b5-a444-46f4-98b6-997531cad4e8" containerName="neutron-db-sync" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.069756 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c95690b5-a444-46f4-98b6-997531cad4e8" containerName="neutron-db-sync" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.071302 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.098160 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-khsc8"] Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.198189 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.198244 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cbx6\" (UniqueName: \"kubernetes.io/projected/ff8b2583-de18-4e08-9161-668b8e950f51-kube-api-access-8cbx6\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.198764 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.198828 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.198892 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.198938 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-config\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.215931 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54644995b8-6dmq7"] Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.232341 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.236546 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.238463 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.238659 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-4q8dr" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.238693 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.266859 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54644995b8-6dmq7"] Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.300471 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.300931 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cbx6\" (UniqueName: \"kubernetes.io/projected/ff8b2583-de18-4e08-9161-668b8e950f51-kube-api-access-8cbx6\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.301057 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.301135 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.301248 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.301329 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-config\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.302574 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.302654 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.302754 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.304158 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.304493 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-config\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.335964 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cbx6\" (UniqueName: \"kubernetes.io/projected/ff8b2583-de18-4e08-9161-668b8e950f51-kube-api-access-8cbx6\") pod \"dnsmasq-dns-84b966f6c9-khsc8\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.403440 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-ovndb-tls-certs\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.403735 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-config\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.403790 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-combined-ca-bundle\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.403862 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t7cm\" (UniqueName: \"kubernetes.io/projected/0a01ce20-c1b5-4208-bb2a-1fbbde607700-kube-api-access-9t7cm\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.403927 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-httpd-config\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.415116 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.506014 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-httpd-config\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.506096 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-ovndb-tls-certs\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.506148 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-config\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.506192 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-combined-ca-bundle\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.506230 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t7cm\" (UniqueName: \"kubernetes.io/projected/0a01ce20-c1b5-4208-bb2a-1fbbde607700-kube-api-access-9t7cm\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.511091 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-combined-ca-bundle\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.511430 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-config\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.512872 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-httpd-config\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.513899 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-ovndb-tls-certs\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.526909 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t7cm\" (UniqueName: \"kubernetes.io/projected/0a01ce20-c1b5-4208-bb2a-1fbbde607700-kube-api-access-9t7cm\") pod \"neutron-54644995b8-6dmq7\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:52 crc kubenswrapper[4788]: I1010 15:05:52.598421 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.223347 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-995b87f75-ghpz6"] Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.225198 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.230422 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.239204 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.244014 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-995b87f75-ghpz6"] Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.365116 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-ovndb-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.365294 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-internal-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.365364 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-config\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.365896 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-combined-ca-bundle\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.365959 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxfb6\" (UniqueName: \"kubernetes.io/projected/e96b96e3-982f-4e27-97a6-4e076fc8ec40-kube-api-access-lxfb6\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.366125 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-httpd-config\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.366228 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-public-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.468929 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-ovndb-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.469006 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-internal-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.469044 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-config\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.469088 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-combined-ca-bundle\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.469118 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxfb6\" (UniqueName: \"kubernetes.io/projected/e96b96e3-982f-4e27-97a6-4e076fc8ec40-kube-api-access-lxfb6\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.469199 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-httpd-config\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.469229 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-public-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.478638 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-combined-ca-bundle\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.479409 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-httpd-config\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.480443 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-ovndb-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.481350 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-config\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.486498 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-public-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.495883 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-internal-tls-certs\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.499713 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxfb6\" (UniqueName: \"kubernetes.io/projected/e96b96e3-982f-4e27-97a6-4e076fc8ec40-kube-api-access-lxfb6\") pod \"neutron-995b87f75-ghpz6\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:55 crc kubenswrapper[4788]: I1010 15:05:55.551077 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.357977 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.509299 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvfrj\" (UniqueName: \"kubernetes.io/projected/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-kube-api-access-hvfrj\") pod \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.509405 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-db-sync-config-data\") pod \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.509452 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-combined-ca-bundle\") pod \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\" (UID: \"1f6f73b5-459a-4490-bac4-426d8b3f2b9e\") " Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.518254 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-kube-api-access-hvfrj" (OuterVolumeSpecName: "kube-api-access-hvfrj") pod "1f6f73b5-459a-4490-bac4-426d8b3f2b9e" (UID: "1f6f73b5-459a-4490-bac4-426d8b3f2b9e"). InnerVolumeSpecName "kube-api-access-hvfrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.518861 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1f6f73b5-459a-4490-bac4-426d8b3f2b9e" (UID: "1f6f73b5-459a-4490-bac4-426d8b3f2b9e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.545166 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f6f73b5-459a-4490-bac4-426d8b3f2b9e" (UID: "1f6f73b5-459a-4490-bac4-426d8b3f2b9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.612203 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvfrj\" (UniqueName: \"kubernetes.io/projected/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-kube-api-access-hvfrj\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.612234 4788 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:57 crc kubenswrapper[4788]: I1010 15:05:57.612244 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6f73b5-459a-4490-bac4-426d8b3f2b9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.000498 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sbnwc" event={"ID":"1f6f73b5-459a-4490-bac4-426d8b3f2b9e","Type":"ContainerDied","Data":"fd48b60312edf8cab58c4402b0cc1b46ede9e593913db08ca4abb692eb4bd9e8"} Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.000827 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd48b60312edf8cab58c4402b0cc1b46ede9e593913db08ca4abb692eb4bd9e8" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.000889 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sbnwc" Oct 10 15:05:58 crc kubenswrapper[4788]: E1010 15:05:58.170863 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" Oct 10 15:05:58 crc kubenswrapper[4788]: E1010 15:05:58.247540 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f6f73b5_459a_4490_bac4_426d8b3f2b9e.slice/crio-fd48b60312edf8cab58c4402b0cc1b46ede9e593913db08ca4abb692eb4bd9e8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f6f73b5_459a_4490_bac4_426d8b3f2b9e.slice\": RecentStats: unable to find data in memory cache]" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.330701 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-khsc8"] Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.538167 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-995b87f75-ghpz6"] Oct 10 15:05:58 crc kubenswrapper[4788]: W1010 15:05:58.539022 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode96b96e3_982f_4e27_97a6_4e076fc8ec40.slice/crio-45f023f4d80ddb824b612b101bda21e1b042bacbc9a3fc698146cc9b7f763e15 WatchSource:0}: Error finding container 45f023f4d80ddb824b612b101bda21e1b042bacbc9a3fc698146cc9b7f763e15: Status 404 returned error can't find the container with id 45f023f4d80ddb824b612b101bda21e1b042bacbc9a3fc698146cc9b7f763e15 Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.706398 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-657c889cbb-d8d5h"] Oct 10 15:05:58 crc kubenswrapper[4788]: E1010 15:05:58.706831 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f6f73b5-459a-4490-bac4-426d8b3f2b9e" containerName="barbican-db-sync" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.706850 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f6f73b5-459a-4490-bac4-426d8b3f2b9e" containerName="barbican-db-sync" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.707065 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f6f73b5-459a-4490-bac4-426d8b3f2b9e" containerName="barbican-db-sync" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.708026 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.710635 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.716708 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9ghf5" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.723050 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.728473 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-657c889cbb-d8d5h"] Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.786362 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5b795d78cf-6sr7n"] Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.791649 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.795172 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.840480 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b795d78cf-6sr7n"] Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.859241 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.859319 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085a2986-2c31-42f5-95c2-55f2414e5dfc-logs\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.859364 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2kqn\" (UniqueName: \"kubernetes.io/projected/085a2986-2c31-42f5-95c2-55f2414e5dfc-kube-api-access-m2kqn\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.859410 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data-custom\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.859481 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-combined-ca-bundle\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.906917 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-khsc8"] Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.935966 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x2hwn"] Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.939042 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.950584 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x2hwn"] Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962624 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data-custom\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962697 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962740 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zbvb\" (UniqueName: \"kubernetes.io/projected/505a880b-6b48-498d-aad2-b30f20d089b6-kube-api-access-4zbvb\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962782 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a880b-6b48-498d-aad2-b30f20d089b6-logs\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962819 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-combined-ca-bundle\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962844 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-combined-ca-bundle\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962881 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962928 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085a2986-2c31-42f5-95c2-55f2414e5dfc-logs\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962949 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data-custom\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.962992 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2kqn\" (UniqueName: \"kubernetes.io/projected/085a2986-2c31-42f5-95c2-55f2414e5dfc-kube-api-access-m2kqn\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.964083 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085a2986-2c31-42f5-95c2-55f2414e5dfc-logs\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.973547 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.984699 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-combined-ca-bundle\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.986700 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data-custom\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:58 crc kubenswrapper[4788]: I1010 15:05:58.988090 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2kqn\" (UniqueName: \"kubernetes.io/projected/085a2986-2c31-42f5-95c2-55f2414e5dfc-kube-api-access-m2kqn\") pod \"barbican-keystone-listener-657c889cbb-d8d5h\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.025357 4788 generic.go:334] "Generic (PLEG): container finished" podID="ff8b2583-de18-4e08-9161-668b8e950f51" containerID="d038abdd70b24fc511e3578f56aa495e62a972fe28d1cfd93935f077c3a6f290" exitCode=0 Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.025459 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" event={"ID":"ff8b2583-de18-4e08-9161-668b8e950f51","Type":"ContainerDied","Data":"d038abdd70b24fc511e3578f56aa495e62a972fe28d1cfd93935f077c3a6f290"} Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.025499 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" event={"ID":"ff8b2583-de18-4e08-9161-668b8e950f51","Type":"ContainerStarted","Data":"992ac93790da174cdf9c66d1563ce2c414c165c8d92d67b30ddbb89e1c30e960"} Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.028274 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-995b87f75-ghpz6" event={"ID":"e96b96e3-982f-4e27-97a6-4e076fc8ec40","Type":"ContainerStarted","Data":"2861b006539119d480308e351b33bc584593a678f96dce7d7c0ab944bb039ad4"} Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.028363 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-995b87f75-ghpz6" event={"ID":"e96b96e3-982f-4e27-97a6-4e076fc8ec40","Type":"ContainerStarted","Data":"45f023f4d80ddb824b612b101bda21e1b042bacbc9a3fc698146cc9b7f763e15"} Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.037124 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-d97566b58-kftdw"] Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.039256 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.041430 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.051609 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d97566b58-kftdw"] Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.056335 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82934fb8-b2e0-4d17-b071-e06e747515f4","Type":"ContainerStarted","Data":"a39f5358541468b172ce1b2d14be325d374226c5cb18b04385a6e40974271560"} Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.056523 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="ceilometer-central-agent" containerID="cri-o://cd905b3981551c017fee3cce5d78db27acb3a006b0a289d22d1980255e28acb8" gracePeriod=30 Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.056692 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.056769 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="proxy-httpd" containerID="cri-o://a39f5358541468b172ce1b2d14be325d374226c5cb18b04385a6e40974271560" gracePeriod=30 Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.056839 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="ceilometer-notification-agent" containerID="cri-o://7ecafbd5dc24dad827ac28f9bc73bf2383bcffe7b45d5665ac712e4299074544" gracePeriod=30 Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064536 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064579 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zbvb\" (UniqueName: \"kubernetes.io/projected/505a880b-6b48-498d-aad2-b30f20d089b6-kube-api-access-4zbvb\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064610 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a880b-6b48-498d-aad2-b30f20d089b6-logs\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064629 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6zp2\" (UniqueName: \"kubernetes.io/projected/a2579a67-a700-4de8-a1d7-0b7a0f45549d-kube-api-access-v6zp2\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064656 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064683 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-combined-ca-bundle\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064718 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-config\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064756 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data-custom\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064777 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064800 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.064825 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.067122 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a880b-6b48-498d-aad2-b30f20d089b6-logs\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.085215 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data-custom\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.085549 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.089525 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.096950 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zbvb\" (UniqueName: \"kubernetes.io/projected/505a880b-6b48-498d-aad2-b30f20d089b6-kube-api-access-4zbvb\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.100873 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-combined-ca-bundle\") pod \"barbican-worker-5b795d78cf-6sr7n\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.130965 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.170878 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-logs\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.170983 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6zp2\" (UniqueName: \"kubernetes.io/projected/a2579a67-a700-4de8-a1d7-0b7a0f45549d-kube-api-access-v6zp2\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.171033 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.171091 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f7gt\" (UniqueName: \"kubernetes.io/projected/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-kube-api-access-6f7gt\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.173336 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.174435 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-config\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.175086 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.175162 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data-custom\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.175281 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-combined-ca-bundle\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.175345 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.175423 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.175515 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.177218 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.177402 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-config\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.178634 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.178920 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.189717 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6zp2\" (UniqueName: \"kubernetes.io/projected/a2579a67-a700-4de8-a1d7-0b7a0f45549d-kube-api-access-v6zp2\") pod \"dnsmasq-dns-75c8ddd69c-x2hwn\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.279469 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.279956 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data-custom\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.280002 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-combined-ca-bundle\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.280119 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-logs\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.280210 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f7gt\" (UniqueName: \"kubernetes.io/projected/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-kube-api-access-6f7gt\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.282206 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-logs\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.288527 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data-custom\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.289555 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.289766 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-combined-ca-bundle\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.299160 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.307696 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f7gt\" (UniqueName: \"kubernetes.io/projected/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-kube-api-access-6f7gt\") pod \"barbican-api-d97566b58-kftdw\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.375891 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.409783 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.409844 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.423089 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54644995b8-6dmq7"] Oct 10 15:05:59 crc kubenswrapper[4788]: W1010 15:05:59.455193 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a01ce20_c1b5_4208_bb2a_1fbbde607700.slice/crio-84545a00bdc37d5b38b882b3860f7032ecbd89a29e7c5015f5242c2792918e07 WatchSource:0}: Error finding container 84545a00bdc37d5b38b882b3860f7032ecbd89a29e7c5015f5242c2792918e07: Status 404 returned error can't find the container with id 84545a00bdc37d5b38b882b3860f7032ecbd89a29e7c5015f5242c2792918e07 Oct 10 15:05:59 crc kubenswrapper[4788]: E1010 15:05:59.664557 4788 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 10 15:05:59 crc kubenswrapper[4788]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/ff8b2583-de18-4e08-9161-668b8e950f51/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 10 15:05:59 crc kubenswrapper[4788]: > podSandboxID="992ac93790da174cdf9c66d1563ce2c414c165c8d92d67b30ddbb89e1c30e960" Oct 10 15:05:59 crc kubenswrapper[4788]: E1010 15:05:59.665070 4788 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 10 15:05:59 crc kubenswrapper[4788]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n574hdbh5ddhffh5c4h5dbh5bfh8bh75h584h97h557h69h55fhbch66dh94h5fdh9fh56fh584h5bbhb5h75hcbh667h77h577h5bhf9hf4hb4q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8cbx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-84b966f6c9-khsc8_openstack(ff8b2583-de18-4e08-9161-668b8e950f51): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/ff8b2583-de18-4e08-9161-668b8e950f51/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 10 15:05:59 crc kubenswrapper[4788]: > logger="UnhandledError" Oct 10 15:05:59 crc kubenswrapper[4788]: E1010 15:05:59.668497 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/ff8b2583-de18-4e08-9161-668b8e950f51/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" podUID="ff8b2583-de18-4e08-9161-668b8e950f51" Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.683000 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-657c889cbb-d8d5h"] Oct 10 15:05:59 crc kubenswrapper[4788]: W1010 15:05:59.707730 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod085a2986_2c31_42f5_95c2_55f2414e5dfc.slice/crio-3ca552bdec7a2eb28d25fc5b5aacd090a4e5b6635f4e9de4433c0f42afa4322a WatchSource:0}: Error finding container 3ca552bdec7a2eb28d25fc5b5aacd090a4e5b6635f4e9de4433c0f42afa4322a: Status 404 returned error can't find the container with id 3ca552bdec7a2eb28d25fc5b5aacd090a4e5b6635f4e9de4433c0f42afa4322a Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.821967 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5b795d78cf-6sr7n"] Oct 10 15:05:59 crc kubenswrapper[4788]: W1010 15:05:59.827689 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod505a880b_6b48_498d_aad2_b30f20d089b6.slice/crio-0731faf896e6f1b658e930bf659b512d62919816ddbf1dfae62248a5cc7acabb WatchSource:0}: Error finding container 0731faf896e6f1b658e930bf659b512d62919816ddbf1dfae62248a5cc7acabb: Status 404 returned error can't find the container with id 0731faf896e6f1b658e930bf659b512d62919816ddbf1dfae62248a5cc7acabb Oct 10 15:05:59 crc kubenswrapper[4788]: I1010 15:05:59.925280 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x2hwn"] Oct 10 15:05:59 crc kubenswrapper[4788]: W1010 15:05:59.935366 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2579a67_a700_4de8_a1d7_0b7a0f45549d.slice/crio-522f2ec408e4594799078a10eb7c88577da4b9b0c50f1af2d774f5f4ab2ad2b5 WatchSource:0}: Error finding container 522f2ec408e4594799078a10eb7c88577da4b9b0c50f1af2d774f5f4ab2ad2b5: Status 404 returned error can't find the container with id 522f2ec408e4594799078a10eb7c88577da4b9b0c50f1af2d774f5f4ab2ad2b5 Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.017850 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d97566b58-kftdw"] Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.068962 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b795d78cf-6sr7n" event={"ID":"505a880b-6b48-498d-aad2-b30f20d089b6","Type":"ContainerStarted","Data":"0731faf896e6f1b658e930bf659b512d62919816ddbf1dfae62248a5cc7acabb"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.071322 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" event={"ID":"a2579a67-a700-4de8-a1d7-0b7a0f45549d","Type":"ContainerStarted","Data":"522f2ec408e4594799078a10eb7c88577da4b9b0c50f1af2d774f5f4ab2ad2b5"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.074374 4788 generic.go:334] "Generic (PLEG): container finished" podID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerID="a39f5358541468b172ce1b2d14be325d374226c5cb18b04385a6e40974271560" exitCode=0 Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.074401 4788 generic.go:334] "Generic (PLEG): container finished" podID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerID="cd905b3981551c017fee3cce5d78db27acb3a006b0a289d22d1980255e28acb8" exitCode=0 Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.074430 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82934fb8-b2e0-4d17-b071-e06e747515f4","Type":"ContainerDied","Data":"a39f5358541468b172ce1b2d14be325d374226c5cb18b04385a6e40974271560"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.074449 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82934fb8-b2e0-4d17-b071-e06e747515f4","Type":"ContainerDied","Data":"cd905b3981551c017fee3cce5d78db27acb3a006b0a289d22d1980255e28acb8"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.076407 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54644995b8-6dmq7" event={"ID":"0a01ce20-c1b5-4208-bb2a-1fbbde607700","Type":"ContainerStarted","Data":"84c28b0499e2a2334685726b3f07e67a51a663809bb94fd0bd87f3a1d89eaddf"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.076434 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54644995b8-6dmq7" event={"ID":"0a01ce20-c1b5-4208-bb2a-1fbbde607700","Type":"ContainerStarted","Data":"de9fa25c2085c639d4eb048cfdfc58620ce3f84bde12fd37936b9ab56e125d8b"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.076445 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54644995b8-6dmq7" event={"ID":"0a01ce20-c1b5-4208-bb2a-1fbbde607700","Type":"ContainerStarted","Data":"84545a00bdc37d5b38b882b3860f7032ecbd89a29e7c5015f5242c2792918e07"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.076589 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.078065 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" event={"ID":"085a2986-2c31-42f5-95c2-55f2414e5dfc","Type":"ContainerStarted","Data":"3ca552bdec7a2eb28d25fc5b5aacd090a4e5b6635f4e9de4433c0f42afa4322a"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.082642 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-995b87f75-ghpz6" event={"ID":"e96b96e3-982f-4e27-97a6-4e076fc8ec40","Type":"ContainerStarted","Data":"5699636a0ee4d24f0612d524e20711c956b0d7f42b60edadc77e819e598ab333"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.083713 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.085407 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d97566b58-kftdw" event={"ID":"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a","Type":"ContainerStarted","Data":"40ad0204db09d522a4504a11af37a46b9ef5255a466bc5037c394c24d6080e9d"} Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.100797 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54644995b8-6dmq7" podStartSLOduration=8.100770994 podStartE2EDuration="8.100770994s" podCreationTimestamp="2025-10-10 15:05:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:00.094112924 +0000 UTC m=+1262.543828482" watchObservedRunningTime="2025-10-10 15:06:00.100770994 +0000 UTC m=+1262.550486542" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.136384 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-995b87f75-ghpz6" podStartSLOduration=5.136359016 podStartE2EDuration="5.136359016s" podCreationTimestamp="2025-10-10 15:05:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:00.130048195 +0000 UTC m=+1262.579763743" watchObservedRunningTime="2025-10-10 15:06:00.136359016 +0000 UTC m=+1262.586074564" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.630463 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.722456 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-config\") pod \"ff8b2583-de18-4e08-9161-668b8e950f51\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.722927 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-sb\") pod \"ff8b2583-de18-4e08-9161-668b8e950f51\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.722998 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-nb\") pod \"ff8b2583-de18-4e08-9161-668b8e950f51\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.723072 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-svc\") pod \"ff8b2583-de18-4e08-9161-668b8e950f51\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.723122 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-swift-storage-0\") pod \"ff8b2583-de18-4e08-9161-668b8e950f51\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.723301 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cbx6\" (UniqueName: \"kubernetes.io/projected/ff8b2583-de18-4e08-9161-668b8e950f51-kube-api-access-8cbx6\") pod \"ff8b2583-de18-4e08-9161-668b8e950f51\" (UID: \"ff8b2583-de18-4e08-9161-668b8e950f51\") " Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.731162 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff8b2583-de18-4e08-9161-668b8e950f51-kube-api-access-8cbx6" (OuterVolumeSpecName: "kube-api-access-8cbx6") pod "ff8b2583-de18-4e08-9161-668b8e950f51" (UID: "ff8b2583-de18-4e08-9161-668b8e950f51"). InnerVolumeSpecName "kube-api-access-8cbx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.774368 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-config" (OuterVolumeSpecName: "config") pod "ff8b2583-de18-4e08-9161-668b8e950f51" (UID: "ff8b2583-de18-4e08-9161-668b8e950f51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.787076 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff8b2583-de18-4e08-9161-668b8e950f51" (UID: "ff8b2583-de18-4e08-9161-668b8e950f51"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.792459 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ff8b2583-de18-4e08-9161-668b8e950f51" (UID: "ff8b2583-de18-4e08-9161-668b8e950f51"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.799801 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff8b2583-de18-4e08-9161-668b8e950f51" (UID: "ff8b2583-de18-4e08-9161-668b8e950f51"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.815793 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff8b2583-de18-4e08-9161-668b8e950f51" (UID: "ff8b2583-de18-4e08-9161-668b8e950f51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.825965 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.826006 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.826019 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.826030 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.826040 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cbx6\" (UniqueName: \"kubernetes.io/projected/ff8b2583-de18-4e08-9161-668b8e950f51-kube-api-access-8cbx6\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:00 crc kubenswrapper[4788]: I1010 15:06:00.826050 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff8b2583-de18-4e08-9161-668b8e950f51-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.101875 4788 generic.go:334] "Generic (PLEG): container finished" podID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerID="0c9a53e76dab791347fc370ae287a1bd6f29f34d29dc9b93531f624b95a7cddc" exitCode=0 Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.101978 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" event={"ID":"a2579a67-a700-4de8-a1d7-0b7a0f45549d","Type":"ContainerDied","Data":"0c9a53e76dab791347fc370ae287a1bd6f29f34d29dc9b93531f624b95a7cddc"} Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.109010 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" event={"ID":"ff8b2583-de18-4e08-9161-668b8e950f51","Type":"ContainerDied","Data":"992ac93790da174cdf9c66d1563ce2c414c165c8d92d67b30ddbb89e1c30e960"} Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.109092 4788 scope.go:117] "RemoveContainer" containerID="d038abdd70b24fc511e3578f56aa495e62a972fe28d1cfd93935f077c3a6f290" Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.109318 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-khsc8" Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.116266 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d97566b58-kftdw" event={"ID":"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a","Type":"ContainerStarted","Data":"86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea"} Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.116395 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d97566b58-kftdw" event={"ID":"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a","Type":"ContainerStarted","Data":"fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3"} Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.117594 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.117734 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.170686 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-d97566b58-kftdw" podStartSLOduration=2.170666123 podStartE2EDuration="2.170666123s" podCreationTimestamp="2025-10-10 15:05:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:01.160245211 +0000 UTC m=+1263.609960759" watchObservedRunningTime="2025-10-10 15:06:01.170666123 +0000 UTC m=+1263.620381661" Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.271471 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-khsc8"] Oct 10 15:06:01 crc kubenswrapper[4788]: I1010 15:06:01.281900 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-khsc8"] Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.127511 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" event={"ID":"a2579a67-a700-4de8-a1d7-0b7a0f45549d","Type":"ContainerStarted","Data":"b6986090a904d3120dd5750e9bba2cb6ff60a1f15d0b1573dfbccd0321c04562"} Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.127991 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.154232 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" podStartSLOduration=4.154212639 podStartE2EDuration="4.154212639s" podCreationTimestamp="2025-10-10 15:05:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:02.148866294 +0000 UTC m=+1264.598581842" watchObservedRunningTime="2025-10-10 15:06:02.154212639 +0000 UTC m=+1264.603928177" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.252575 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff8b2583-de18-4e08-9161-668b8e950f51" path="/var/lib/kubelet/pods/ff8b2583-de18-4e08-9161-668b8e950f51/volumes" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.676246 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-bc8db69bb-s2ddx"] Oct 10 15:06:02 crc kubenswrapper[4788]: E1010 15:06:02.677239 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff8b2583-de18-4e08-9161-668b8e950f51" containerName="init" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.677320 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff8b2583-de18-4e08-9161-668b8e950f51" containerName="init" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.677891 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff8b2583-de18-4e08-9161-668b8e950f51" containerName="init" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.679133 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.681677 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.682051 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.701866 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-bc8db69bb-s2ddx"] Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.848159 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-combined-ca-bundle\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.848205 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data-custom\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.848248 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2qw6\" (UniqueName: \"kubernetes.io/projected/7b604164-1660-4f3a-929b-cf6f59b66823-kube-api-access-n2qw6\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.848264 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-public-tls-certs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.848746 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-internal-tls-certs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.848854 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.848890 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b604164-1660-4f3a-929b-cf6f59b66823-logs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.951896 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-internal-tls-certs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.951961 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.951985 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b604164-1660-4f3a-929b-cf6f59b66823-logs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.952036 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-combined-ca-bundle\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.952068 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data-custom\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.952112 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2qw6\" (UniqueName: \"kubernetes.io/projected/7b604164-1660-4f3a-929b-cf6f59b66823-kube-api-access-n2qw6\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.952132 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-public-tls-certs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.955061 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b604164-1660-4f3a-929b-cf6f59b66823-logs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.959553 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-combined-ca-bundle\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.960100 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-public-tls-certs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.960169 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data-custom\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.960538 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-internal-tls-certs\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.962418 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:02 crc kubenswrapper[4788]: I1010 15:06:02.990294 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2qw6\" (UniqueName: \"kubernetes.io/projected/7b604164-1660-4f3a-929b-cf6f59b66823-kube-api-access-n2qw6\") pod \"barbican-api-bc8db69bb-s2ddx\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:03 crc kubenswrapper[4788]: I1010 15:06:02.999975 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.185071 4788 generic.go:334] "Generic (PLEG): container finished" podID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerID="7ecafbd5dc24dad827ac28f9bc73bf2383bcffe7b45d5665ac712e4299074544" exitCode=0 Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.185124 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82934fb8-b2e0-4d17-b071-e06e747515f4","Type":"ContainerDied","Data":"7ecafbd5dc24dad827ac28f9bc73bf2383bcffe7b45d5665ac712e4299074544"} Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.185830 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"82934fb8-b2e0-4d17-b071-e06e747515f4","Type":"ContainerDied","Data":"ae09df5e4acbc2dd0cff3489cdc76943478bfc338ab99b35c526e1ccdf22edd5"} Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.185845 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae09df5e4acbc2dd0cff3489cdc76943478bfc338ab99b35c526e1ccdf22edd5" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.369488 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.487096 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-config-data\") pod \"82934fb8-b2e0-4d17-b071-e06e747515f4\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.487612 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klvsv\" (UniqueName: \"kubernetes.io/projected/82934fb8-b2e0-4d17-b071-e06e747515f4-kube-api-access-klvsv\") pod \"82934fb8-b2e0-4d17-b071-e06e747515f4\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.487737 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-log-httpd\") pod \"82934fb8-b2e0-4d17-b071-e06e747515f4\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.487826 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-sg-core-conf-yaml\") pod \"82934fb8-b2e0-4d17-b071-e06e747515f4\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.488032 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-combined-ca-bundle\") pod \"82934fb8-b2e0-4d17-b071-e06e747515f4\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.488330 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "82934fb8-b2e0-4d17-b071-e06e747515f4" (UID: "82934fb8-b2e0-4d17-b071-e06e747515f4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.488130 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-scripts\") pod \"82934fb8-b2e0-4d17-b071-e06e747515f4\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.490119 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-run-httpd\") pod \"82934fb8-b2e0-4d17-b071-e06e747515f4\" (UID: \"82934fb8-b2e0-4d17-b071-e06e747515f4\") " Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.490806 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.491400 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "82934fb8-b2e0-4d17-b071-e06e747515f4" (UID: "82934fb8-b2e0-4d17-b071-e06e747515f4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.493527 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-scripts" (OuterVolumeSpecName: "scripts") pod "82934fb8-b2e0-4d17-b071-e06e747515f4" (UID: "82934fb8-b2e0-4d17-b071-e06e747515f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.494827 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82934fb8-b2e0-4d17-b071-e06e747515f4-kube-api-access-klvsv" (OuterVolumeSpecName: "kube-api-access-klvsv") pod "82934fb8-b2e0-4d17-b071-e06e747515f4" (UID: "82934fb8-b2e0-4d17-b071-e06e747515f4"). InnerVolumeSpecName "kube-api-access-klvsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.496795 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "82934fb8-b2e0-4d17-b071-e06e747515f4" (UID: "82934fb8-b2e0-4d17-b071-e06e747515f4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.529767 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-bc8db69bb-s2ddx"] Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.592479 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.592509 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.592518 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/82934fb8-b2e0-4d17-b071-e06e747515f4-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.592528 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klvsv\" (UniqueName: \"kubernetes.io/projected/82934fb8-b2e0-4d17-b071-e06e747515f4-kube-api-access-klvsv\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.596032 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82934fb8-b2e0-4d17-b071-e06e747515f4" (UID: "82934fb8-b2e0-4d17-b071-e06e747515f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.628244 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-config-data" (OuterVolumeSpecName: "config-data") pod "82934fb8-b2e0-4d17-b071-e06e747515f4" (UID: "82934fb8-b2e0-4d17-b071-e06e747515f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.694482 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:04 crc kubenswrapper[4788]: I1010 15:06:04.694513 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82934fb8-b2e0-4d17-b071-e06e747515f4-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.202535 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b795d78cf-6sr7n" event={"ID":"505a880b-6b48-498d-aad2-b30f20d089b6","Type":"ContainerStarted","Data":"b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee"} Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.202854 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b795d78cf-6sr7n" event={"ID":"505a880b-6b48-498d-aad2-b30f20d089b6","Type":"ContainerStarted","Data":"3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950"} Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.211178 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" event={"ID":"085a2986-2c31-42f5-95c2-55f2414e5dfc","Type":"ContainerStarted","Data":"705a07a6c50b14bb6069ed9b5c58612e9e5861f6620fb1c72d2da2cfd85e0e7b"} Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.211236 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" event={"ID":"085a2986-2c31-42f5-95c2-55f2414e5dfc","Type":"ContainerStarted","Data":"3c499468e5604cdfd26560a7f04ecbdd2ff30447627e047d1fc5e0e0d7e4cd3b"} Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.214771 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.214798 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bc8db69bb-s2ddx" event={"ID":"7b604164-1660-4f3a-929b-cf6f59b66823","Type":"ContainerStarted","Data":"f1acd019454438f9f318884feab766dfda302a7f2270489b71e0fc3660adf204"} Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.214878 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bc8db69bb-s2ddx" event={"ID":"7b604164-1660-4f3a-929b-cf6f59b66823","Type":"ContainerStarted","Data":"441c3dc286f891485385a909d54daf00c136c609bb2e0e40814c55b244268d27"} Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.214896 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bc8db69bb-s2ddx" event={"ID":"7b604164-1660-4f3a-929b-cf6f59b66823","Type":"ContainerStarted","Data":"7af9a6e24bee3d2c197114ed28318653e221a3c48158feaad3ed7eb762d58f09"} Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.215611 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.215688 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.229548 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5b795d78cf-6sr7n" podStartSLOduration=3.000104866 podStartE2EDuration="7.229523923s" podCreationTimestamp="2025-10-10 15:05:58 +0000 UTC" firstStartedPulling="2025-10-10 15:05:59.830666109 +0000 UTC m=+1262.280381657" lastFinishedPulling="2025-10-10 15:06:04.060085166 +0000 UTC m=+1266.509800714" observedRunningTime="2025-10-10 15:06:05.227329484 +0000 UTC m=+1267.677045032" watchObservedRunningTime="2025-10-10 15:06:05.229523923 +0000 UTC m=+1267.679239471" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.275169 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-bc8db69bb-s2ddx" podStartSLOduration=3.275125255 podStartE2EDuration="3.275125255s" podCreationTimestamp="2025-10-10 15:06:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:05.251279121 +0000 UTC m=+1267.701012769" watchObservedRunningTime="2025-10-10 15:06:05.275125255 +0000 UTC m=+1267.724840793" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.297240 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" podStartSLOduration=2.9538068859999997 podStartE2EDuration="7.297217582s" podCreationTimestamp="2025-10-10 15:05:58 +0000 UTC" firstStartedPulling="2025-10-10 15:05:59.714825071 +0000 UTC m=+1262.164540619" lastFinishedPulling="2025-10-10 15:06:04.058235767 +0000 UTC m=+1266.507951315" observedRunningTime="2025-10-10 15:06:05.273383078 +0000 UTC m=+1267.723098646" watchObservedRunningTime="2025-10-10 15:06:05.297217582 +0000 UTC m=+1267.746933120" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.364930 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.370702 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.378006 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:05 crc kubenswrapper[4788]: E1010 15:06:05.378575 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="proxy-httpd" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.378591 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="proxy-httpd" Oct 10 15:06:05 crc kubenswrapper[4788]: E1010 15:06:05.378632 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="ceilometer-notification-agent" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.378639 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="ceilometer-notification-agent" Oct 10 15:06:05 crc kubenswrapper[4788]: E1010 15:06:05.378659 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="ceilometer-central-agent" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.378666 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="ceilometer-central-agent" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.378848 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="proxy-httpd" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.378882 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="ceilometer-central-agent" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.378893 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" containerName="ceilometer-notification-agent" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.380890 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.394878 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.395178 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.409313 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.518189 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bczzw\" (UniqueName: \"kubernetes.io/projected/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-kube-api-access-bczzw\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.518683 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.518963 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-scripts\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.519075 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-log-httpd\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.519156 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-config-data\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.519198 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-run-httpd\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.519247 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.621977 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-scripts\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.622066 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-log-httpd\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.622103 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-config-data\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.622126 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-run-httpd\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.622233 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.622572 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-log-httpd\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.623674 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bczzw\" (UniqueName: \"kubernetes.io/projected/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-kube-api-access-bczzw\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.623771 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.629880 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-run-httpd\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.646062 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-config-data\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.646062 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bczzw\" (UniqueName: \"kubernetes.io/projected/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-kube-api-access-bczzw\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.649716 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.650212 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-scripts\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.659703 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " pod="openstack/ceilometer-0" Oct 10 15:06:05 crc kubenswrapper[4788]: I1010 15:06:05.724967 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:06 crc kubenswrapper[4788]: I1010 15:06:06.209267 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:06 crc kubenswrapper[4788]: I1010 15:06:06.225933 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerStarted","Data":"58b440433ed2eece160b1b6d8bef661642a87093c9b81f83020df41e7c018df8"} Oct 10 15:06:06 crc kubenswrapper[4788]: I1010 15:06:06.228731 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cnk7r" event={"ID":"63cf2515-0dd5-4f1f-b70e-08a095284b53","Type":"ContainerStarted","Data":"8e6c023e337efd4e7921b84ae53a248be93d4aba90a3ae0477341ae598b257a9"} Oct 10 15:06:06 crc kubenswrapper[4788]: I1010 15:06:06.267177 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-cnk7r" podStartSLOduration=7.985091293 podStartE2EDuration="45.267134168s" podCreationTimestamp="2025-10-10 15:05:21 +0000 UTC" firstStartedPulling="2025-10-10 15:05:27.36703271 +0000 UTC m=+1229.816748258" lastFinishedPulling="2025-10-10 15:06:04.649075585 +0000 UTC m=+1267.098791133" observedRunningTime="2025-10-10 15:06:06.260269212 +0000 UTC m=+1268.709984760" watchObservedRunningTime="2025-10-10 15:06:06.267134168 +0000 UTC m=+1268.716849896" Oct 10 15:06:06 crc kubenswrapper[4788]: I1010 15:06:06.295797 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82934fb8-b2e0-4d17-b071-e06e747515f4" path="/var/lib/kubelet/pods/82934fb8-b2e0-4d17-b071-e06e747515f4/volumes" Oct 10 15:06:07 crc kubenswrapper[4788]: I1010 15:06:07.239116 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerStarted","Data":"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed"} Oct 10 15:06:07 crc kubenswrapper[4788]: I1010 15:06:07.518348 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:06:07 crc kubenswrapper[4788]: I1010 15:06:07.524660 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:06:08 crc kubenswrapper[4788]: I1010 15:06:08.259061 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerStarted","Data":"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5"} Oct 10 15:06:09 crc kubenswrapper[4788]: I1010 15:06:09.269750 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerStarted","Data":"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49"} Oct 10 15:06:09 crc kubenswrapper[4788]: I1010 15:06:09.301898 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:06:09 crc kubenswrapper[4788]: I1010 15:06:09.378189 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9wvx7"] Oct 10 15:06:09 crc kubenswrapper[4788]: I1010 15:06:09.378658 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" podUID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" containerName="dnsmasq-dns" containerID="cri-o://42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9" gracePeriod=10 Oct 10 15:06:09 crc kubenswrapper[4788]: I1010 15:06:09.985019 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.133089 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhpxz\" (UniqueName: \"kubernetes.io/projected/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-kube-api-access-nhpxz\") pod \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.133459 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-swift-storage-0\") pod \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.133521 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-svc\") pod \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.133616 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-nb\") pod \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.133831 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-sb\") pod \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.133850 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-config\") pod \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\" (UID: \"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3\") " Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.141521 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-kube-api-access-nhpxz" (OuterVolumeSpecName: "kube-api-access-nhpxz") pod "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" (UID: "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3"). InnerVolumeSpecName "kube-api-access-nhpxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.235905 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhpxz\" (UniqueName: \"kubernetes.io/projected/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-kube-api-access-nhpxz\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.236710 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-config" (OuterVolumeSpecName: "config") pod "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" (UID: "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.281828 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" (UID: "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.287248 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" (UID: "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.291299 4788 generic.go:334] "Generic (PLEG): container finished" podID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" containerID="42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9" exitCode=0 Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.291341 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" event={"ID":"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3","Type":"ContainerDied","Data":"42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9"} Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.291370 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" event={"ID":"3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3","Type":"ContainerDied","Data":"c1039dab291f6e117ae135158db3ad708e176c16435b3a138a2cc488a6d37718"} Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.291390 4788 scope.go:117] "RemoveContainer" containerID="42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.291542 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-9wvx7" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.296225 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.308349 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" (UID: "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.312959 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" (UID: "3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.330406 4788 scope.go:117] "RemoveContainer" containerID="bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.351943 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.351978 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.351989 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.352000 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.352012 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.407921 4788 scope.go:117] "RemoveContainer" containerID="42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9" Oct 10 15:06:10 crc kubenswrapper[4788]: E1010 15:06:10.408593 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9\": container with ID starting with 42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9 not found: ID does not exist" containerID="42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.408631 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9"} err="failed to get container status \"42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9\": rpc error: code = NotFound desc = could not find container \"42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9\": container with ID starting with 42ff4b94453bfd3204a601a2ee2c25d7c5a944a083787629f28fc2443a1d75b9 not found: ID does not exist" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.408658 4788 scope.go:117] "RemoveContainer" containerID="bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e" Oct 10 15:06:10 crc kubenswrapper[4788]: E1010 15:06:10.409102 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e\": container with ID starting with bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e not found: ID does not exist" containerID="bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.409161 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e"} err="failed to get container status \"bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e\": rpc error: code = NotFound desc = could not find container \"bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e\": container with ID starting with bc7bf951bac1ff165aa0afd11ced21bb13505170ab87511d6619dc02b9e0898e not found: ID does not exist" Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.624337 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9wvx7"] Oct 10 15:06:10 crc kubenswrapper[4788]: I1010 15:06:10.632410 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-9wvx7"] Oct 10 15:06:11 crc kubenswrapper[4788]: I1010 15:06:11.302625 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerStarted","Data":"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e"} Oct 10 15:06:11 crc kubenswrapper[4788]: I1010 15:06:11.302798 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 15:06:11 crc kubenswrapper[4788]: I1010 15:06:11.307405 4788 generic.go:334] "Generic (PLEG): container finished" podID="63cf2515-0dd5-4f1f-b70e-08a095284b53" containerID="8e6c023e337efd4e7921b84ae53a248be93d4aba90a3ae0477341ae598b257a9" exitCode=0 Oct 10 15:06:11 crc kubenswrapper[4788]: I1010 15:06:11.307459 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cnk7r" event={"ID":"63cf2515-0dd5-4f1f-b70e-08a095284b53","Type":"ContainerDied","Data":"8e6c023e337efd4e7921b84ae53a248be93d4aba90a3ae0477341ae598b257a9"} Oct 10 15:06:11 crc kubenswrapper[4788]: I1010 15:06:11.331770 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5425299580000003 podStartE2EDuration="6.331743765s" podCreationTimestamp="2025-10-10 15:06:05 +0000 UTC" firstStartedPulling="2025-10-10 15:06:06.202257536 +0000 UTC m=+1268.651973084" lastFinishedPulling="2025-10-10 15:06:09.991471343 +0000 UTC m=+1272.441186891" observedRunningTime="2025-10-10 15:06:11.329426912 +0000 UTC m=+1273.779142460" watchObservedRunningTime="2025-10-10 15:06:11.331743765 +0000 UTC m=+1273.781459313" Oct 10 15:06:11 crc kubenswrapper[4788]: I1010 15:06:11.365779 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:06:11 crc kubenswrapper[4788]: I1010 15:06:11.731889 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.248273 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" path="/var/lib/kubelet/pods/3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3/volumes" Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.743072 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.931232 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-scripts\") pod \"63cf2515-0dd5-4f1f-b70e-08a095284b53\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.931322 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7wm4\" (UniqueName: \"kubernetes.io/projected/63cf2515-0dd5-4f1f-b70e-08a095284b53-kube-api-access-p7wm4\") pod \"63cf2515-0dd5-4f1f-b70e-08a095284b53\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.931384 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-db-sync-config-data\") pod \"63cf2515-0dd5-4f1f-b70e-08a095284b53\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.931441 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-combined-ca-bundle\") pod \"63cf2515-0dd5-4f1f-b70e-08a095284b53\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.931534 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63cf2515-0dd5-4f1f-b70e-08a095284b53-etc-machine-id\") pod \"63cf2515-0dd5-4f1f-b70e-08a095284b53\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.931583 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-config-data\") pod \"63cf2515-0dd5-4f1f-b70e-08a095284b53\" (UID: \"63cf2515-0dd5-4f1f-b70e-08a095284b53\") " Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.939290 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "63cf2515-0dd5-4f1f-b70e-08a095284b53" (UID: "63cf2515-0dd5-4f1f-b70e-08a095284b53"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.940423 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63cf2515-0dd5-4f1f-b70e-08a095284b53-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "63cf2515-0dd5-4f1f-b70e-08a095284b53" (UID: "63cf2515-0dd5-4f1f-b70e-08a095284b53"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.943240 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-scripts" (OuterVolumeSpecName: "scripts") pod "63cf2515-0dd5-4f1f-b70e-08a095284b53" (UID: "63cf2515-0dd5-4f1f-b70e-08a095284b53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.954809 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63cf2515-0dd5-4f1f-b70e-08a095284b53-kube-api-access-p7wm4" (OuterVolumeSpecName: "kube-api-access-p7wm4") pod "63cf2515-0dd5-4f1f-b70e-08a095284b53" (UID: "63cf2515-0dd5-4f1f-b70e-08a095284b53"). InnerVolumeSpecName "kube-api-access-p7wm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.979525 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63cf2515-0dd5-4f1f-b70e-08a095284b53" (UID: "63cf2515-0dd5-4f1f-b70e-08a095284b53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:12 crc kubenswrapper[4788]: I1010 15:06:12.996824 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-config-data" (OuterVolumeSpecName: "config-data") pod "63cf2515-0dd5-4f1f-b70e-08a095284b53" (UID: "63cf2515-0dd5-4f1f-b70e-08a095284b53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.033477 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.033515 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7wm4\" (UniqueName: \"kubernetes.io/projected/63cf2515-0dd5-4f1f-b70e-08a095284b53-kube-api-access-p7wm4\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.033528 4788 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.033537 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.033546 4788 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63cf2515-0dd5-4f1f-b70e-08a095284b53-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.033555 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63cf2515-0dd5-4f1f-b70e-08a095284b53-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.327553 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cnk7r" event={"ID":"63cf2515-0dd5-4f1f-b70e-08a095284b53","Type":"ContainerDied","Data":"c9f56d84d8e9eded0c1fa3db8d05aa26bcf64626e2c71f61c6ea757de96fd379"} Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.327602 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9f56d84d8e9eded0c1fa3db8d05aa26bcf64626e2c71f61c6ea757de96fd379" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.327659 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cnk7r" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.589489 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:13 crc kubenswrapper[4788]: E1010 15:06:13.589981 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63cf2515-0dd5-4f1f-b70e-08a095284b53" containerName="cinder-db-sync" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.589996 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="63cf2515-0dd5-4f1f-b70e-08a095284b53" containerName="cinder-db-sync" Oct 10 15:06:13 crc kubenswrapper[4788]: E1010 15:06:13.590019 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" containerName="dnsmasq-dns" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.590025 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" containerName="dnsmasq-dns" Oct 10 15:06:13 crc kubenswrapper[4788]: E1010 15:06:13.590046 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" containerName="init" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.590053 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" containerName="init" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.590271 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ac65ff4-c322-4c5c-85f6-b01b33e9c4b3" containerName="dnsmasq-dns" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.590285 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="63cf2515-0dd5-4f1f-b70e-08a095284b53" containerName="cinder-db-sync" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.596888 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.600114 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.600399 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2fdbx" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.600744 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.600864 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.611908 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.709603 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5bj6f"] Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.722515 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.726721 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5bj6f"] Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.751376 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx5zg\" (UniqueName: \"kubernetes.io/projected/8044d012-63a7-4d09-9b28-9d45c6a49367-kube-api-access-gx5zg\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.751529 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8044d012-63a7-4d09-9b28-9d45c6a49367-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.751583 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.751636 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.751668 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-scripts\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.751747 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855114 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8044d012-63a7-4d09-9b28-9d45c6a49367-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855203 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-config\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855233 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855259 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855304 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855329 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-scripts\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855354 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855387 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws82x\" (UniqueName: \"kubernetes.io/projected/8f684fc3-b2c2-454a-b70e-5a19f9520710-kube-api-access-ws82x\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855412 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855435 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855460 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-svc\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855483 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx5zg\" (UniqueName: \"kubernetes.io/projected/8044d012-63a7-4d09-9b28-9d45c6a49367-kube-api-access-gx5zg\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.855876 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8044d012-63a7-4d09-9b28-9d45c6a49367-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.892631 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.913712 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.923799 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx5zg\" (UniqueName: \"kubernetes.io/projected/8044d012-63a7-4d09-9b28-9d45c6a49367-kube-api-access-gx5zg\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.924645 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.925542 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.927407 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-scripts\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.929507 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.929902 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data\") pod \"cinder-scheduler-0\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.932684 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.959831 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-config\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.959888 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.959958 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.959991 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws82x\" (UniqueName: \"kubernetes.io/projected/8f684fc3-b2c2-454a-b70e-5a19f9520710-kube-api-access-ws82x\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.960017 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.960044 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-svc\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.962210 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.983010 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-svc\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.983390 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-config\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.984420 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.984974 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:13 crc kubenswrapper[4788]: I1010 15:06:13.985581 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.024817 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws82x\" (UniqueName: \"kubernetes.io/projected/8f684fc3-b2c2-454a-b70e-5a19f9520710-kube-api-access-ws82x\") pod \"dnsmasq-dns-5784cf869f-5bj6f\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.065362 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.065415 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data-custom\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.065437 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-scripts\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.065462 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.065504 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59jjz\" (UniqueName: \"kubernetes.io/projected/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-kube-api-access-59jjz\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.065559 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.065589 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-logs\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.073542 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.169929 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-logs\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.170353 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-logs\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.170512 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.170578 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data-custom\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.170596 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-scripts\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.170619 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.170654 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59jjz\" (UniqueName: \"kubernetes.io/projected/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-kube-api-access-59jjz\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.170707 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.170550 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.178985 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-scripts\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.182834 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data-custom\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.183499 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.183946 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.189913 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59jjz\" (UniqueName: \"kubernetes.io/projected/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-kube-api-access-59jjz\") pod \"cinder-api-0\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.314897 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.574729 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:14 crc kubenswrapper[4788]: W1010 15:06:14.589303 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8044d012_63a7_4d09_9b28_9d45c6a49367.slice/crio-90759e0de3144ec41fce7e6ac0f155e8f23e7587974c9159e8060ea42b84ae13 WatchSource:0}: Error finding container 90759e0de3144ec41fce7e6ac0f155e8f23e7587974c9159e8060ea42b84ae13: Status 404 returned error can't find the container with id 90759e0de3144ec41fce7e6ac0f155e8f23e7587974c9159e8060ea42b84ae13 Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.728366 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.730841 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.743078 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.744473 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.746574 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zgbbl" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.763878 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.808587 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5bj6f"] Oct 10 15:06:14 crc kubenswrapper[4788]: W1010 15:06:14.815710 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f684fc3_b2c2_454a_b70e_5a19f9520710.slice/crio-51aea72c498a2a95988c0dc5418c3d7f01427762774eaf668f2fb9c27c3851b3 WatchSource:0}: Error finding container 51aea72c498a2a95988c0dc5418c3d7f01427762774eaf668f2fb9c27c3851b3: Status 404 returned error can't find the container with id 51aea72c498a2a95988c0dc5418c3d7f01427762774eaf668f2fb9c27c3851b3 Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.889544 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.889633 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbjfc\" (UniqueName: \"kubernetes.io/projected/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-kube-api-access-kbjfc\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.889654 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config-secret\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.889694 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.928828 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.942652 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.982616 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.991685 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbjfc\" (UniqueName: \"kubernetes.io/projected/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-kube-api-access-kbjfc\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.991760 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config-secret\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.991837 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.992034 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:14 crc kubenswrapper[4788]: I1010 15:06:14.993093 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.104962 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-d97566b58-kftdw"] Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.113347 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-d97566b58-kftdw" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerName="barbican-api-log" containerID="cri-o://fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3" gracePeriod=30 Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.113921 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-d97566b58-kftdw" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerName="barbican-api" containerID="cri-o://86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea" gracePeriod=30 Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.163439 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 10 15:06:15 crc kubenswrapper[4788]: E1010 15:06:15.167346 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-kbjfc openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="2aac1bd6-7032-4911-a6ed-5c15b24b8c9c" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.182112 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.248190 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.249430 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.268181 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.361813 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.370877 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config-secret\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.381108 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e7ccf50d-1227-42b5-aaf5-3cb3471faad4","Type":"ContainerStarted","Data":"3479250bfe5ea8d501ffa37482b5bef834f1df8367d26d3b374377f675c899f5"} Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.390750 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbjfc\" (UniqueName: \"kubernetes.io/projected/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-kube-api-access-kbjfc\") pod \"openstackclient\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.394957 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" event={"ID":"8f684fc3-b2c2-454a-b70e-5a19f9520710","Type":"ContainerStarted","Data":"51aea72c498a2a95988c0dc5418c3d7f01427762774eaf668f2fb9c27c3851b3"} Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.401616 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbjwq\" (UniqueName: \"kubernetes.io/projected/d84e4226-ac3f-4f6f-870b-c795e206854e-kube-api-access-zbjwq\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.401670 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.401743 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.401777 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config-secret\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.404645 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8044d012-63a7-4d09-9b28-9d45c6a49367","Type":"ContainerStarted","Data":"90759e0de3144ec41fce7e6ac0f155e8f23e7587974c9159e8060ea42b84ae13"} Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.404791 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.416730 4788 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2aac1bd6-7032-4911-a6ed-5c15b24b8c9c" podUID="d84e4226-ac3f-4f6f-870b-c795e206854e" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.507911 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbjwq\" (UniqueName: \"kubernetes.io/projected/d84e4226-ac3f-4f6f-870b-c795e206854e-kube-api-access-zbjwq\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.507961 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.509186 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.513361 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.513452 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config-secret\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.531255 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.534694 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbjwq\" (UniqueName: \"kubernetes.io/projected/d84e4226-ac3f-4f6f-870b-c795e206854e-kube-api-access-zbjwq\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.534850 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config-secret\") pod \"openstackclient\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.544343 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.721792 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config\") pod \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.722566 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "2aac1bd6-7032-4911-a6ed-5c15b24b8c9c" (UID: "2aac1bd6-7032-4911-a6ed-5c15b24b8c9c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.722892 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbjfc\" (UniqueName: \"kubernetes.io/projected/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-kube-api-access-kbjfc\") pod \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.722934 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config-secret\") pod \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.722989 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-combined-ca-bundle\") pod \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\" (UID: \"2aac1bd6-7032-4911-a6ed-5c15b24b8c9c\") " Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.723381 4788 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.727855 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-kube-api-access-kbjfc" (OuterVolumeSpecName: "kube-api-access-kbjfc") pod "2aac1bd6-7032-4911-a6ed-5c15b24b8c9c" (UID: "2aac1bd6-7032-4911-a6ed-5c15b24b8c9c"). InnerVolumeSpecName "kube-api-access-kbjfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.728509 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "2aac1bd6-7032-4911-a6ed-5c15b24b8c9c" (UID: "2aac1bd6-7032-4911-a6ed-5c15b24b8c9c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.728579 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2aac1bd6-7032-4911-a6ed-5c15b24b8c9c" (UID: "2aac1bd6-7032-4911-a6ed-5c15b24b8c9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.821077 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.825285 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbjfc\" (UniqueName: \"kubernetes.io/projected/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-kube-api-access-kbjfc\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.825315 4788 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:15 crc kubenswrapper[4788]: I1010 15:06:15.825326 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.293406 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aac1bd6-7032-4911-a6ed-5c15b24b8c9c" path="/var/lib/kubelet/pods/2aac1bd6-7032-4911-a6ed-5c15b24b8c9c/volumes" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.299316 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-87b66898f-smc95"] Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.301391 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.307610 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.307708 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.329325 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.355615 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-87b66898f-smc95"] Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.460820 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-public-tls-certs\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.461229 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-config-data\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.461472 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpzpv\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-kube-api-access-gpzpv\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.461770 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-run-httpd\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.461911 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-internal-tls-certs\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.462051 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-etc-swift\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.462248 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-log-httpd\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.462255 4788 generic.go:334] "Generic (PLEG): container finished" podID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerID="fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3" exitCode=143 Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.462360 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d97566b58-kftdw" event={"ID":"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a","Type":"ContainerDied","Data":"fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3"} Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.462695 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-combined-ca-bundle\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.469374 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.478542 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e7ccf50d-1227-42b5-aaf5-3cb3471faad4","Type":"ContainerStarted","Data":"3f97b52d47c54bb90b4b7160365c3a49e2c886ca8a65ee3949f187e17ec1123b"} Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.485379 4788 generic.go:334] "Generic (PLEG): container finished" podID="8f684fc3-b2c2-454a-b70e-5a19f9520710" containerID="824819b6460de36355916687ebd695df87d0acd9246071dee316aeadb206bf16" exitCode=0 Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.485768 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.486345 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" event={"ID":"8f684fc3-b2c2-454a-b70e-5a19f9520710","Type":"ContainerDied","Data":"824819b6460de36355916687ebd695df87d0acd9246071dee316aeadb206bf16"} Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.540940 4788 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="2aac1bd6-7032-4911-a6ed-5c15b24b8c9c" podUID="d84e4226-ac3f-4f6f-870b-c795e206854e" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.564927 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpzpv\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-kube-api-access-gpzpv\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.565117 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-run-httpd\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.565252 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-internal-tls-certs\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.565349 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-etc-swift\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.565432 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-log-httpd\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.565542 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-combined-ca-bundle\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.565618 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-public-tls-certs\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.565697 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-config-data\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.567606 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-log-httpd\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.567867 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-run-httpd\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.570097 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-config-data\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.573622 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-combined-ca-bundle\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.577547 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-etc-swift\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.578113 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-internal-tls-certs\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.580754 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-public-tls-certs\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.584732 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpzpv\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-kube-api-access-gpzpv\") pod \"swift-proxy-87b66898f-smc95\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.677940 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:16 crc kubenswrapper[4788]: I1010 15:06:16.866813 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.466862 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-87b66898f-smc95"] Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.489905 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.490218 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="ceilometer-central-agent" containerID="cri-o://704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed" gracePeriod=30 Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.490628 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="proxy-httpd" containerID="cri-o://046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e" gracePeriod=30 Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.490677 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="sg-core" containerID="cri-o://a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49" gracePeriod=30 Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.490713 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="ceilometer-notification-agent" containerID="cri-o://284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5" gracePeriod=30 Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.521544 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-87b66898f-smc95" event={"ID":"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f","Type":"ContainerStarted","Data":"048ed87aed0d749cfb7ffc903044db9c0bd9beac32a54479ee9f835b1b965c82"} Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.527738 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" event={"ID":"8f684fc3-b2c2-454a-b70e-5a19f9520710","Type":"ContainerStarted","Data":"a2a9700b17de8390730939a608c45fe4b3b60e9b1e67395aceadf7ac2576f14f"} Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.529333 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.542980 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8044d012-63a7-4d09-9b28-9d45c6a49367","Type":"ContainerStarted","Data":"6072b842a5e2fd46f1b5bef829db2c0e3a6ef3b6c68e465a327bcee97c048730"} Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.550864 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d84e4226-ac3f-4f6f-870b-c795e206854e","Type":"ContainerStarted","Data":"1e5d9d3696357f385c69d49f868d71a91f88e61fd48d8291594a1761b0450e1e"} Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.560247 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" podStartSLOduration=4.560226268 podStartE2EDuration="4.560226268s" podCreationTimestamp="2025-10-10 15:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:17.558794329 +0000 UTC m=+1280.008509877" watchObservedRunningTime="2025-10-10 15:06:17.560226268 +0000 UTC m=+1280.009941816" Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.566486 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e7ccf50d-1227-42b5-aaf5-3cb3471faad4","Type":"ContainerStarted","Data":"75bc72d77ca51763e142eda1335a09a768c99c2bf9ade041fdec8f17c2aa8354"} Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.566692 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerName="cinder-api-log" containerID="cri-o://3f97b52d47c54bb90b4b7160365c3a49e2c886ca8a65ee3949f187e17ec1123b" gracePeriod=30 Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.566837 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.567327 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerName="cinder-api" containerID="cri-o://75bc72d77ca51763e142eda1335a09a768c99c2bf9ade041fdec8f17c2aa8354" gracePeriod=30 Oct 10 15:06:17 crc kubenswrapper[4788]: I1010 15:06:17.599671 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.599653202 podStartE2EDuration="4.599653202s" podCreationTimestamp="2025-10-10 15:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:17.591185334 +0000 UTC m=+1280.040900882" watchObservedRunningTime="2025-10-10 15:06:17.599653202 +0000 UTC m=+1280.049368750" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.466821 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.595496 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8044d012-63a7-4d09-9b28-9d45c6a49367","Type":"ContainerStarted","Data":"c8430737887e50131054ce04a3980073899d78e2ce0a1dd9c67a4d27052e9ebc"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.607335 4788 generic.go:334] "Generic (PLEG): container finished" podID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerID="3f97b52d47c54bb90b4b7160365c3a49e2c886ca8a65ee3949f187e17ec1123b" exitCode=143 Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.607405 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e7ccf50d-1227-42b5-aaf5-3cb3471faad4","Type":"ContainerDied","Data":"3f97b52d47c54bb90b4b7160365c3a49e2c886ca8a65ee3949f187e17ec1123b"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.612363 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-87b66898f-smc95" event={"ID":"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f","Type":"ContainerStarted","Data":"87673bff375f480e6dff20a3a6b396e073584e22ecfa29abe8788960477c7c73"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.612418 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-87b66898f-smc95" event={"ID":"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f","Type":"ContainerStarted","Data":"fd696a35a1e5c7857218ceaedea76b07e06bb96d506d78970a1c5927133e9d90"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.613277 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.613303 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.622371 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-sg-core-conf-yaml\") pod \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.622543 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-config-data\") pod \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.622597 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-run-httpd\") pod \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.622615 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-scripts\") pod \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.622658 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-log-httpd\") pod \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.622677 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-combined-ca-bundle\") pod \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.622709 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bczzw\" (UniqueName: \"kubernetes.io/projected/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-kube-api-access-bczzw\") pod \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\" (UID: \"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727\") " Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.623342 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" (UID: "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.624185 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.624212 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" (UID: "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.628004 4788 generic.go:334] "Generic (PLEG): container finished" podID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerID="046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e" exitCode=0 Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.628074 4788 generic.go:334] "Generic (PLEG): container finished" podID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerID="a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49" exitCode=2 Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.628087 4788 generic.go:334] "Generic (PLEG): container finished" podID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerID="284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5" exitCode=0 Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.628101 4788 generic.go:334] "Generic (PLEG): container finished" podID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerID="704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed" exitCode=0 Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.628861 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.629288 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerDied","Data":"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.629398 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerDied","Data":"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.629417 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerDied","Data":"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.629428 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerDied","Data":"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.629440 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6d02d53-5ea4-4944-a7e6-28a4e5f5b727","Type":"ContainerDied","Data":"58b440433ed2eece160b1b6d8bef661642a87093c9b81f83020df41e7c018df8"} Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.629435 4788 scope.go:117] "RemoveContainer" containerID="046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.630459 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-scripts" (OuterVolumeSpecName: "scripts") pod "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" (UID: "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.643339 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-kube-api-access-bczzw" (OuterVolumeSpecName: "kube-api-access-bczzw") pod "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" (UID: "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727"). InnerVolumeSpecName "kube-api-access-bczzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.647525 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.473685471 podStartE2EDuration="5.647497806s" podCreationTimestamp="2025-10-10 15:06:13 +0000 UTC" firstStartedPulling="2025-10-10 15:06:14.59627072 +0000 UTC m=+1277.045986268" lastFinishedPulling="2025-10-10 15:06:15.770083055 +0000 UTC m=+1278.219798603" observedRunningTime="2025-10-10 15:06:18.614758291 +0000 UTC m=+1281.064473839" watchObservedRunningTime="2025-10-10 15:06:18.647497806 +0000 UTC m=+1281.097213354" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.652586 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-87b66898f-smc95" podStartSLOduration=2.6525704919999997 podStartE2EDuration="2.652570492s" podCreationTimestamp="2025-10-10 15:06:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:18.634701819 +0000 UTC m=+1281.084417367" watchObservedRunningTime="2025-10-10 15:06:18.652570492 +0000 UTC m=+1281.102286040" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.715799 4788 scope.go:117] "RemoveContainer" containerID="a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.731097 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" (UID: "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.740868 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.741098 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.741214 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bczzw\" (UniqueName: \"kubernetes.io/projected/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-kube-api-access-bczzw\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.741290 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.770953 4788 scope.go:117] "RemoveContainer" containerID="284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.798133 4788 scope.go:117] "RemoveContainer" containerID="704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.799266 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" (UID: "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.831556 4788 scope.go:117] "RemoveContainer" containerID="046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e" Oct 10 15:06:18 crc kubenswrapper[4788]: E1010 15:06:18.832088 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": container with ID starting with 046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e not found: ID does not exist" containerID="046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.832131 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e"} err="failed to get container status \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": rpc error: code = NotFound desc = could not find container \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": container with ID starting with 046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.832182 4788 scope.go:117] "RemoveContainer" containerID="a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49" Oct 10 15:06:18 crc kubenswrapper[4788]: E1010 15:06:18.832777 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": container with ID starting with a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49 not found: ID does not exist" containerID="a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.832803 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49"} err="failed to get container status \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": rpc error: code = NotFound desc = could not find container \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": container with ID starting with a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49 not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.832817 4788 scope.go:117] "RemoveContainer" containerID="284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5" Oct 10 15:06:18 crc kubenswrapper[4788]: E1010 15:06:18.833644 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": container with ID starting with 284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5 not found: ID does not exist" containerID="284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.833666 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5"} err="failed to get container status \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": rpc error: code = NotFound desc = could not find container \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": container with ID starting with 284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5 not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.833679 4788 scope.go:117] "RemoveContainer" containerID="704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed" Oct 10 15:06:18 crc kubenswrapper[4788]: E1010 15:06:18.834171 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": container with ID starting with 704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed not found: ID does not exist" containerID="704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.834195 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed"} err="failed to get container status \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": rpc error: code = NotFound desc = could not find container \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": container with ID starting with 704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.834208 4788 scope.go:117] "RemoveContainer" containerID="046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.834707 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e"} err="failed to get container status \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": rpc error: code = NotFound desc = could not find container \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": container with ID starting with 046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.834729 4788 scope.go:117] "RemoveContainer" containerID="a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.835276 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49"} err="failed to get container status \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": rpc error: code = NotFound desc = could not find container \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": container with ID starting with a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49 not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.835295 4788 scope.go:117] "RemoveContainer" containerID="284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.835635 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5"} err="failed to get container status \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": rpc error: code = NotFound desc = could not find container \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": container with ID starting with 284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5 not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.835666 4788 scope.go:117] "RemoveContainer" containerID="704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.836108 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed"} err="failed to get container status \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": rpc error: code = NotFound desc = could not find container \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": container with ID starting with 704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.836128 4788 scope.go:117] "RemoveContainer" containerID="046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.836399 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e"} err="failed to get container status \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": rpc error: code = NotFound desc = could not find container \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": container with ID starting with 046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.836434 4788 scope.go:117] "RemoveContainer" containerID="a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.837656 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49"} err="failed to get container status \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": rpc error: code = NotFound desc = could not find container \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": container with ID starting with a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49 not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.837694 4788 scope.go:117] "RemoveContainer" containerID="284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.839476 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5"} err="failed to get container status \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": rpc error: code = NotFound desc = could not find container \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": container with ID starting with 284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5 not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.839527 4788 scope.go:117] "RemoveContainer" containerID="704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.840264 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed"} err="failed to get container status \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": rpc error: code = NotFound desc = could not find container \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": container with ID starting with 704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.840288 4788 scope.go:117] "RemoveContainer" containerID="046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.840759 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e"} err="failed to get container status \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": rpc error: code = NotFound desc = could not find container \"046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e\": container with ID starting with 046d68bed1e728c7ae675a2ed120280b580402bafb0e9e0712dbd1df3d714e9e not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.840781 4788 scope.go:117] "RemoveContainer" containerID="a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.842058 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49"} err="failed to get container status \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": rpc error: code = NotFound desc = could not find container \"a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49\": container with ID starting with a367dd09c2f3388fb86b69fb7e202581da1009b25e08ed0b11e07c4e3671bd49 not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.842085 4788 scope.go:117] "RemoveContainer" containerID="284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.842713 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.843542 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5"} err="failed to get container status \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": rpc error: code = NotFound desc = could not find container \"284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5\": container with ID starting with 284b4adb94eae3973da107e9cdee65a91facb937ccdb3e82304ea043043550f5 not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.843573 4788 scope.go:117] "RemoveContainer" containerID="704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.843796 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-config-data" (OuterVolumeSpecName: "config-data") pod "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" (UID: "a6d02d53-5ea4-4944-a7e6-28a4e5f5b727"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.843963 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed"} err="failed to get container status \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": rpc error: code = NotFound desc = could not find container \"704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed\": container with ID starting with 704e4505579fcc5c83d5df2d31af79c5fda49d38918413d8f092ce89c7e893ed not found: ID does not exist" Oct 10 15:06:18 crc kubenswrapper[4788]: E1010 15:06:18.868516 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf71185c6_5c8c_4a6d_82e8_4d799a9bb63a.slice/crio-86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf71185c6_5c8c_4a6d_82e8_4d799a9bb63a.slice/crio-conmon-86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea.scope\": RecentStats: unable to find data in memory cache]" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.933458 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.946021 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:18 crc kubenswrapper[4788]: I1010 15:06:18.995434 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.013690 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.028945 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:19 crc kubenswrapper[4788]: E1010 15:06:19.029416 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="ceilometer-central-agent" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.029437 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="ceilometer-central-agent" Oct 10 15:06:19 crc kubenswrapper[4788]: E1010 15:06:19.029459 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="ceilometer-notification-agent" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.029466 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="ceilometer-notification-agent" Oct 10 15:06:19 crc kubenswrapper[4788]: E1010 15:06:19.029503 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="sg-core" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.029509 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="sg-core" Oct 10 15:06:19 crc kubenswrapper[4788]: E1010 15:06:19.029516 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="proxy-httpd" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.029524 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="proxy-httpd" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.029710 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="ceilometer-notification-agent" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.029740 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="sg-core" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.029750 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="proxy-httpd" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.029760 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" containerName="ceilometer-central-agent" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.031842 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.034852 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.035086 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.040324 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.152510 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-run-httpd\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.152600 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.152623 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.152668 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-log-httpd\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.152995 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-config-data\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.153083 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrzv7\" (UniqueName: \"kubernetes.io/projected/313e22c1-d5ed-485b-a22d-9d76b36c060e-kube-api-access-zrzv7\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.153162 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-scripts\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.255580 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-scripts\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.256011 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-run-httpd\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.256042 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.256062 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.256081 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-log-httpd\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.256177 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-config-data\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.256233 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrzv7\" (UniqueName: \"kubernetes.io/projected/313e22c1-d5ed-485b-a22d-9d76b36c060e-kube-api-access-zrzv7\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.256949 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-run-httpd\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.257636 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-log-httpd\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.266988 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.268811 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-config-data\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.273036 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-scripts\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.273773 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.285169 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrzv7\" (UniqueName: \"kubernetes.io/projected/313e22c1-d5ed-485b-a22d-9d76b36c060e-kube-api-access-zrzv7\") pod \"ceilometer-0\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.349670 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.355912 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.459092 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-logs\") pod \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.459318 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-combined-ca-bundle\") pod \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.459386 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f7gt\" (UniqueName: \"kubernetes.io/projected/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-kube-api-access-6f7gt\") pod \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.459516 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data-custom\") pod \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.459542 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data\") pod \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\" (UID: \"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a\") " Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.463060 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-logs" (OuterVolumeSpecName: "logs") pod "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" (UID: "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.469336 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-kube-api-access-6f7gt" (OuterVolumeSpecName: "kube-api-access-6f7gt") pod "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" (UID: "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a"). InnerVolumeSpecName "kube-api-access-6f7gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.473289 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" (UID: "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.537292 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" (UID: "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.538894 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data" (OuterVolumeSpecName: "config-data") pod "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" (UID: "f71185c6-5c8c-4a6d-82e8-4d799a9bb63a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.561767 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.561803 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.561815 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f7gt\" (UniqueName: \"kubernetes.io/projected/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-kube-api-access-6f7gt\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.561826 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.561836 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.682455 4788 generic.go:334] "Generic (PLEG): container finished" podID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerID="86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea" exitCode=0 Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.682834 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d97566b58-kftdw" event={"ID":"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a","Type":"ContainerDied","Data":"86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea"} Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.682865 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d97566b58-kftdw" event={"ID":"f71185c6-5c8c-4a6d-82e8-4d799a9bb63a","Type":"ContainerDied","Data":"40ad0204db09d522a4504a11af37a46b9ef5255a466bc5037c394c24d6080e9d"} Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.682882 4788 scope.go:117] "RemoveContainer" containerID="86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.683024 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d97566b58-kftdw" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.734756 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-d97566b58-kftdw"] Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.742026 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-d97566b58-kftdw"] Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.762077 4788 scope.go:117] "RemoveContainer" containerID="fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.819834 4788 scope.go:117] "RemoveContainer" containerID="86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea" Oct 10 15:06:19 crc kubenswrapper[4788]: E1010 15:06:19.820398 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea\": container with ID starting with 86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea not found: ID does not exist" containerID="86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.820447 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea"} err="failed to get container status \"86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea\": rpc error: code = NotFound desc = could not find container \"86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea\": container with ID starting with 86e7cb89d257ca710160f508f686c96cd316fe5dd7c6d16ee32eec8d7abd99ea not found: ID does not exist" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.820496 4788 scope.go:117] "RemoveContainer" containerID="fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3" Oct 10 15:06:19 crc kubenswrapper[4788]: E1010 15:06:19.824528 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3\": container with ID starting with fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3 not found: ID does not exist" containerID="fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3" Oct 10 15:06:19 crc kubenswrapper[4788]: I1010 15:06:19.824553 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3"} err="failed to get container status \"fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3\": rpc error: code = NotFound desc = could not find container \"fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3\": container with ID starting with fb089e83d4650c83382828fc097567d75fc0195b13b84c3b980142c0e44a0cd3 not found: ID does not exist" Oct 10 15:06:20 crc kubenswrapper[4788]: I1010 15:06:20.040777 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:20 crc kubenswrapper[4788]: I1010 15:06:20.251381 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6d02d53-5ea4-4944-a7e6-28a4e5f5b727" path="/var/lib/kubelet/pods/a6d02d53-5ea4-4944-a7e6-28a4e5f5b727/volumes" Oct 10 15:06:20 crc kubenswrapper[4788]: I1010 15:06:20.256686 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" path="/var/lib/kubelet/pods/f71185c6-5c8c-4a6d-82e8-4d799a9bb63a/volumes" Oct 10 15:06:20 crc kubenswrapper[4788]: I1010 15:06:20.704837 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerStarted","Data":"ba0c63908bd6e05784cc0396573a2ebcd04ae40bbe258cc295f3e7174c8753fa"} Oct 10 15:06:21 crc kubenswrapper[4788]: I1010 15:06:21.721050 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerStarted","Data":"c3a92a869ed5908e17e2d01bba28698ca41954f1fc612d3ee7e536963a8ff050"} Oct 10 15:06:21 crc kubenswrapper[4788]: I1010 15:06:21.721403 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerStarted","Data":"67cef068036c5b37dba000bedcb9bf0ff6c33de9d830345baad07b8a59071afa"} Oct 10 15:06:22 crc kubenswrapper[4788]: I1010 15:06:22.611534 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:06:22 crc kubenswrapper[4788]: I1010 15:06:22.739577 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerStarted","Data":"09298c4c2ba5002454492fc39933dfc6adedae8c83aa2a613ea9f1d4ec67c96d"} Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.076466 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.143938 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x2hwn"] Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.144499 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerName="dnsmasq-dns" containerID="cri-o://b6986090a904d3120dd5750e9bba2cb6ff60a1f15d0b1573dfbccd0321c04562" gracePeriod=10 Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.277313 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.301623 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.347771 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.776738 4788 generic.go:334] "Generic (PLEG): container finished" podID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerID="b6986090a904d3120dd5750e9bba2cb6ff60a1f15d0b1573dfbccd0321c04562" exitCode=0 Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.776995 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerName="cinder-scheduler" containerID="cri-o://6072b842a5e2fd46f1b5bef829db2c0e3a6ef3b6c68e465a327bcee97c048730" gracePeriod=30 Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.777293 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" event={"ID":"a2579a67-a700-4de8-a1d7-0b7a0f45549d","Type":"ContainerDied","Data":"b6986090a904d3120dd5750e9bba2cb6ff60a1f15d0b1573dfbccd0321c04562"} Oct 10 15:06:24 crc kubenswrapper[4788]: I1010 15:06:24.777670 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerName="probe" containerID="cri-o://c8430737887e50131054ce04a3980073899d78e2ce0a1dd9c67a4d27052e9ebc" gracePeriod=30 Oct 10 15:06:25 crc kubenswrapper[4788]: I1010 15:06:25.089692 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:25 crc kubenswrapper[4788]: I1010 15:06:25.569575 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:06:25 crc kubenswrapper[4788]: I1010 15:06:25.641728 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54644995b8-6dmq7"] Oct 10 15:06:25 crc kubenswrapper[4788]: I1010 15:06:25.641975 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54644995b8-6dmq7" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerName="neutron-api" containerID="cri-o://de9fa25c2085c639d4eb048cfdfc58620ce3f84bde12fd37936b9ab56e125d8b" gracePeriod=30 Oct 10 15:06:25 crc kubenswrapper[4788]: I1010 15:06:25.642092 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54644995b8-6dmq7" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerName="neutron-httpd" containerID="cri-o://84c28b0499e2a2334685726b3f07e67a51a663809bb94fd0bd87f3a1d89eaddf" gracePeriod=30 Oct 10 15:06:26 crc kubenswrapper[4788]: I1010 15:06:26.687104 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:26 crc kubenswrapper[4788]: I1010 15:06:26.687198 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:06:26 crc kubenswrapper[4788]: I1010 15:06:26.823847 4788 generic.go:334] "Generic (PLEG): container finished" podID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerID="84c28b0499e2a2334685726b3f07e67a51a663809bb94fd0bd87f3a1d89eaddf" exitCode=0 Oct 10 15:06:26 crc kubenswrapper[4788]: I1010 15:06:26.824015 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54644995b8-6dmq7" event={"ID":"0a01ce20-c1b5-4208-bb2a-1fbbde607700","Type":"ContainerDied","Data":"84c28b0499e2a2334685726b3f07e67a51a663809bb94fd0bd87f3a1d89eaddf"} Oct 10 15:06:26 crc kubenswrapper[4788]: I1010 15:06:26.826954 4788 generic.go:334] "Generic (PLEG): container finished" podID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerID="c8430737887e50131054ce04a3980073899d78e2ce0a1dd9c67a4d27052e9ebc" exitCode=0 Oct 10 15:06:26 crc kubenswrapper[4788]: I1010 15:06:26.826999 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8044d012-63a7-4d09-9b28-9d45c6a49367","Type":"ContainerDied","Data":"c8430737887e50131054ce04a3980073899d78e2ce0a1dd9c67a4d27052e9ebc"} Oct 10 15:06:26 crc kubenswrapper[4788]: I1010 15:06:26.977639 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 10 15:06:27 crc kubenswrapper[4788]: I1010 15:06:27.845442 4788 generic.go:334] "Generic (PLEG): container finished" podID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerID="de9fa25c2085c639d4eb048cfdfc58620ce3f84bde12fd37936b9ab56e125d8b" exitCode=0 Oct 10 15:06:27 crc kubenswrapper[4788]: I1010 15:06:27.845507 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54644995b8-6dmq7" event={"ID":"0a01ce20-c1b5-4208-bb2a-1fbbde607700","Type":"ContainerDied","Data":"de9fa25c2085c639d4eb048cfdfc58620ce3f84bde12fd37936b9ab56e125d8b"} Oct 10 15:06:29 crc kubenswrapper[4788]: E1010 15:06:29.186046 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8044d012_63a7_4d09_9b28_9d45c6a49367.slice/crio-conmon-6072b842a5e2fd46f1b5bef829db2c0e3a6ef3b6c68e465a327bcee97c048730.scope\": RecentStats: unable to find data in memory cache]" Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.300258 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.406290 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.406351 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.406396 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.407163 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e260922adcdc34acddb0d03ca667c174167bee6681d51128f1688ce0f11d52b"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.407232 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://8e260922adcdc34acddb0d03ca667c174167bee6681d51128f1688ce0f11d52b" gracePeriod=600 Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.875266 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="8e260922adcdc34acddb0d03ca667c174167bee6681d51128f1688ce0f11d52b" exitCode=0 Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.876276 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"8e260922adcdc34acddb0d03ca667c174167bee6681d51128f1688ce0f11d52b"} Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.876686 4788 scope.go:117] "RemoveContainer" containerID="9bf2bdef66313687037ed74946bc83398b51a4a5e0533a877fae35897091d486" Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.897900 4788 generic.go:334] "Generic (PLEG): container finished" podID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerID="6072b842a5e2fd46f1b5bef829db2c0e3a6ef3b6c68e465a327bcee97c048730" exitCode=0 Oct 10 15:06:29 crc kubenswrapper[4788]: I1010 15:06:29.897958 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8044d012-63a7-4d09-9b28-9d45c6a49367","Type":"ContainerDied","Data":"6072b842a5e2fd46f1b5bef829db2c0e3a6ef3b6c68e465a327bcee97c048730"} Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.075716 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.107740 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.119414 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data\") pod \"8044d012-63a7-4d09-9b28-9d45c6a49367\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.119507 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-svc\") pod \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.119601 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-sb\") pod \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.119693 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-nb\") pod \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.119721 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data-custom\") pod \"8044d012-63a7-4d09-9b28-9d45c6a49367\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.119775 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-config\") pod \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.119825 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6zp2\" (UniqueName: \"kubernetes.io/projected/a2579a67-a700-4de8-a1d7-0b7a0f45549d-kube-api-access-v6zp2\") pod \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.119969 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx5zg\" (UniqueName: \"kubernetes.io/projected/8044d012-63a7-4d09-9b28-9d45c6a49367-kube-api-access-gx5zg\") pod \"8044d012-63a7-4d09-9b28-9d45c6a49367\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.120004 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-scripts\") pod \"8044d012-63a7-4d09-9b28-9d45c6a49367\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.120344 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-swift-storage-0\") pod \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\" (UID: \"a2579a67-a700-4de8-a1d7-0b7a0f45549d\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.120379 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-combined-ca-bundle\") pod \"8044d012-63a7-4d09-9b28-9d45c6a49367\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.120410 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8044d012-63a7-4d09-9b28-9d45c6a49367-etc-machine-id\") pod \"8044d012-63a7-4d09-9b28-9d45c6a49367\" (UID: \"8044d012-63a7-4d09-9b28-9d45c6a49367\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.120966 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8044d012-63a7-4d09-9b28-9d45c6a49367-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8044d012-63a7-4d09-9b28-9d45c6a49367" (UID: "8044d012-63a7-4d09-9b28-9d45c6a49367"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.125673 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8044d012-63a7-4d09-9b28-9d45c6a49367" (UID: "8044d012-63a7-4d09-9b28-9d45c6a49367"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.129414 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8044d012-63a7-4d09-9b28-9d45c6a49367-kube-api-access-gx5zg" (OuterVolumeSpecName: "kube-api-access-gx5zg") pod "8044d012-63a7-4d09-9b28-9d45c6a49367" (UID: "8044d012-63a7-4d09-9b28-9d45c6a49367"). InnerVolumeSpecName "kube-api-access-gx5zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.131480 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-scripts" (OuterVolumeSpecName: "scripts") pod "8044d012-63a7-4d09-9b28-9d45c6a49367" (UID: "8044d012-63a7-4d09-9b28-9d45c6a49367"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.134594 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2579a67-a700-4de8-a1d7-0b7a0f45549d-kube-api-access-v6zp2" (OuterVolumeSpecName: "kube-api-access-v6zp2") pod "a2579a67-a700-4de8-a1d7-0b7a0f45549d" (UID: "a2579a67-a700-4de8-a1d7-0b7a0f45549d"). InnerVolumeSpecName "kube-api-access-v6zp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.219469 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-config" (OuterVolumeSpecName: "config") pod "a2579a67-a700-4de8-a1d7-0b7a0f45549d" (UID: "a2579a67-a700-4de8-a1d7-0b7a0f45549d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.223926 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx5zg\" (UniqueName: \"kubernetes.io/projected/8044d012-63a7-4d09-9b28-9d45c6a49367-kube-api-access-gx5zg\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.223950 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.223960 4788 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8044d012-63a7-4d09-9b28-9d45c6a49367-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.223971 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.223980 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.223989 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6zp2\" (UniqueName: \"kubernetes.io/projected/a2579a67-a700-4de8-a1d7-0b7a0f45549d-kube-api-access-v6zp2\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.234754 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a2579a67-a700-4de8-a1d7-0b7a0f45549d" (UID: "a2579a67-a700-4de8-a1d7-0b7a0f45549d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.269850 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2579a67-a700-4de8-a1d7-0b7a0f45549d" (UID: "a2579a67-a700-4de8-a1d7-0b7a0f45549d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.271205 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a2579a67-a700-4de8-a1d7-0b7a0f45549d" (UID: "a2579a67-a700-4de8-a1d7-0b7a0f45549d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.283690 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8044d012-63a7-4d09-9b28-9d45c6a49367" (UID: "8044d012-63a7-4d09-9b28-9d45c6a49367"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.297960 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2579a67-a700-4de8-a1d7-0b7a0f45549d" (UID: "a2579a67-a700-4de8-a1d7-0b7a0f45549d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.321654 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.326133 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.326169 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.326180 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.326192 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.326205 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2579a67-a700-4de8-a1d7-0b7a0f45549d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.331234 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data" (OuterVolumeSpecName: "config-data") pod "8044d012-63a7-4d09-9b28-9d45c6a49367" (UID: "8044d012-63a7-4d09-9b28-9d45c6a49367"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.427469 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t7cm\" (UniqueName: \"kubernetes.io/projected/0a01ce20-c1b5-4208-bb2a-1fbbde607700-kube-api-access-9t7cm\") pod \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.427535 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-combined-ca-bundle\") pod \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.427565 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-ovndb-tls-certs\") pod \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.427599 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-config\") pod \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.427642 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-httpd-config\") pod \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\" (UID: \"0a01ce20-c1b5-4208-bb2a-1fbbde607700\") " Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.427880 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8044d012-63a7-4d09-9b28-9d45c6a49367-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.437327 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0a01ce20-c1b5-4208-bb2a-1fbbde607700" (UID: "0a01ce20-c1b5-4208-bb2a-1fbbde607700"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.440352 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a01ce20-c1b5-4208-bb2a-1fbbde607700-kube-api-access-9t7cm" (OuterVolumeSpecName: "kube-api-access-9t7cm") pod "0a01ce20-c1b5-4208-bb2a-1fbbde607700" (UID: "0a01ce20-c1b5-4208-bb2a-1fbbde607700"). InnerVolumeSpecName "kube-api-access-9t7cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.473763 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a01ce20-c1b5-4208-bb2a-1fbbde607700" (UID: "0a01ce20-c1b5-4208-bb2a-1fbbde607700"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.489250 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-config" (OuterVolumeSpecName: "config") pod "0a01ce20-c1b5-4208-bb2a-1fbbde607700" (UID: "0a01ce20-c1b5-4208-bb2a-1fbbde607700"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.507359 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0a01ce20-c1b5-4208-bb2a-1fbbde607700" (UID: "0a01ce20-c1b5-4208-bb2a-1fbbde607700"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.529202 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.529232 4788 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.529243 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.529252 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0a01ce20-c1b5-4208-bb2a-1fbbde607700-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.529262 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t7cm\" (UniqueName: \"kubernetes.io/projected/0a01ce20-c1b5-4208-bb2a-1fbbde607700-kube-api-access-9t7cm\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.910718 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782"} Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.914899 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.914969 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x2hwn" event={"ID":"a2579a67-a700-4de8-a1d7-0b7a0f45549d","Type":"ContainerDied","Data":"522f2ec408e4594799078a10eb7c88577da4b9b0c50f1af2d774f5f4ab2ad2b5"} Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.915072 4788 scope.go:117] "RemoveContainer" containerID="b6986090a904d3120dd5750e9bba2cb6ff60a1f15d0b1573dfbccd0321c04562" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.919169 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerStarted","Data":"fb75102a14192d4f5364f015c3d44a361468374ef6a0e3277df9f9bab699db00"} Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.930828 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="ceilometer-notification-agent" containerID="cri-o://c3a92a869ed5908e17e2d01bba28698ca41954f1fc612d3ee7e536963a8ff050" gracePeriod=30 Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.930830 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.930928 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="proxy-httpd" containerID="cri-o://fb75102a14192d4f5364f015c3d44a361468374ef6a0e3277df9f9bab699db00" gracePeriod=30 Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.930836 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="sg-core" containerID="cri-o://09298c4c2ba5002454492fc39933dfc6adedae8c83aa2a613ea9f1d4ec67c96d" gracePeriod=30 Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.936384 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="ceilometer-central-agent" containerID="cri-o://67cef068036c5b37dba000bedcb9bf0ff6c33de9d830345baad07b8a59071afa" gracePeriod=30 Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.946951 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54644995b8-6dmq7" event={"ID":"0a01ce20-c1b5-4208-bb2a-1fbbde607700","Type":"ContainerDied","Data":"84545a00bdc37d5b38b882b3860f7032ecbd89a29e7c5015f5242c2792918e07"} Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.947173 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54644995b8-6dmq7" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.954180 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8044d012-63a7-4d09-9b28-9d45c6a49367","Type":"ContainerDied","Data":"90759e0de3144ec41fce7e6ac0f155e8f23e7587974c9159e8060ea42b84ae13"} Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.954405 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.971894 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d84e4226-ac3f-4f6f-870b-c795e206854e","Type":"ContainerStarted","Data":"62068765dbbe9a789befff7b25adc8fe43b248f71c27fd2edbc0898452134000"} Oct 10 15:06:30 crc kubenswrapper[4788]: I1010 15:06:30.987773 4788 scope.go:117] "RemoveContainer" containerID="0c9a53e76dab791347fc370ae287a1bd6f29f34d29dc9b93531f624b95a7cddc" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.018022 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.358729648 podStartE2EDuration="13.018000447s" podCreationTimestamp="2025-10-10 15:06:18 +0000 UTC" firstStartedPulling="2025-10-10 15:06:20.060367978 +0000 UTC m=+1282.510083526" lastFinishedPulling="2025-10-10 15:06:29.719638777 +0000 UTC m=+1292.169354325" observedRunningTime="2025-10-10 15:06:30.96592827 +0000 UTC m=+1293.415643818" watchObservedRunningTime="2025-10-10 15:06:31.018000447 +0000 UTC m=+1293.467715995" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.027440 4788 scope.go:117] "RemoveContainer" containerID="84c28b0499e2a2334685726b3f07e67a51a663809bb94fd0bd87f3a1d89eaddf" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.035825 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.8303865950000002 podStartE2EDuration="16.035791266s" podCreationTimestamp="2025-10-10 15:06:15 +0000 UTC" firstStartedPulling="2025-10-10 15:06:16.536349473 +0000 UTC m=+1278.986065021" lastFinishedPulling="2025-10-10 15:06:29.741754154 +0000 UTC m=+1292.191469692" observedRunningTime="2025-10-10 15:06:30.990479743 +0000 UTC m=+1293.440195291" watchObservedRunningTime="2025-10-10 15:06:31.035791266 +0000 UTC m=+1293.485506814" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.064881 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x2hwn"] Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.074892 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x2hwn"] Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.082656 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54644995b8-6dmq7"] Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.090169 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-54644995b8-6dmq7"] Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.113465 4788 scope.go:117] "RemoveContainer" containerID="de9fa25c2085c639d4eb048cfdfc58620ce3f84bde12fd37936b9ab56e125d8b" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.118340 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.126803 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.140516 4788 scope.go:117] "RemoveContainer" containerID="c8430737887e50131054ce04a3980073899d78e2ce0a1dd9c67a4d27052e9ebc" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150018 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:31 crc kubenswrapper[4788]: E1010 15:06:31.150627 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerName="barbican-api" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150659 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerName="barbican-api" Oct 10 15:06:31 crc kubenswrapper[4788]: E1010 15:06:31.150678 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerName="cinder-scheduler" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150688 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerName="cinder-scheduler" Oct 10 15:06:31 crc kubenswrapper[4788]: E1010 15:06:31.150701 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerName="neutron-httpd" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150710 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerName="neutron-httpd" Oct 10 15:06:31 crc kubenswrapper[4788]: E1010 15:06:31.150725 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerName="probe" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150733 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerName="probe" Oct 10 15:06:31 crc kubenswrapper[4788]: E1010 15:06:31.150744 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerName="init" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150751 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerName="init" Oct 10 15:06:31 crc kubenswrapper[4788]: E1010 15:06:31.150767 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerName="barbican-api-log" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150775 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerName="barbican-api-log" Oct 10 15:06:31 crc kubenswrapper[4788]: E1010 15:06:31.150786 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerName="dnsmasq-dns" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150793 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerName="dnsmasq-dns" Oct 10 15:06:31 crc kubenswrapper[4788]: E1010 15:06:31.150808 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerName="neutron-api" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.150817 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerName="neutron-api" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.151089 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerName="barbican-api-log" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.151108 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerName="cinder-scheduler" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.151126 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f71185c6-5c8c-4a6d-82e8-4d799a9bb63a" containerName="barbican-api" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.151154 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerName="neutron-api" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.151170 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" containerName="neutron-httpd" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.151183 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" containerName="dnsmasq-dns" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.151202 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" containerName="probe" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.152575 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.155455 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.169603 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.209984 4788 scope.go:117] "RemoveContainer" containerID="6072b842a5e2fd46f1b5bef829db2c0e3a6ef3b6c68e465a327bcee97c048730" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.254916 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-scripts\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.254974 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.255158 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmvpf\" (UniqueName: \"kubernetes.io/projected/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-kube-api-access-bmvpf\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.255242 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.255605 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.255652 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.357721 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.357770 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.357817 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-scripts\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.357840 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.357921 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.358842 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmvpf\" (UniqueName: \"kubernetes.io/projected/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-kube-api-access-bmvpf\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.358896 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.368358 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-scripts\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.372315 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.374835 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.375049 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.389660 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmvpf\" (UniqueName: \"kubernetes.io/projected/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-kube-api-access-bmvpf\") pod \"cinder-scheduler-0\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " pod="openstack/cinder-scheduler-0" Oct 10 15:06:31 crc kubenswrapper[4788]: I1010 15:06:31.513674 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.025285 4788 generic.go:334] "Generic (PLEG): container finished" podID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerID="fb75102a14192d4f5364f015c3d44a361468374ef6a0e3277df9f9bab699db00" exitCode=0 Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.025854 4788 generic.go:334] "Generic (PLEG): container finished" podID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerID="09298c4c2ba5002454492fc39933dfc6adedae8c83aa2a613ea9f1d4ec67c96d" exitCode=2 Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.025863 4788 generic.go:334] "Generic (PLEG): container finished" podID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerID="c3a92a869ed5908e17e2d01bba28698ca41954f1fc612d3ee7e536963a8ff050" exitCode=0 Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.025871 4788 generic.go:334] "Generic (PLEG): container finished" podID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerID="67cef068036c5b37dba000bedcb9bf0ff6c33de9d830345baad07b8a59071afa" exitCode=0 Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.025923 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerDied","Data":"fb75102a14192d4f5364f015c3d44a361468374ef6a0e3277df9f9bab699db00"} Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.025951 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerDied","Data":"09298c4c2ba5002454492fc39933dfc6adedae8c83aa2a613ea9f1d4ec67c96d"} Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.025964 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerDied","Data":"c3a92a869ed5908e17e2d01bba28698ca41954f1fc612d3ee7e536963a8ff050"} Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.025973 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerDied","Data":"67cef068036c5b37dba000bedcb9bf0ff6c33de9d830345baad07b8a59071afa"} Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.032675 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.176849 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.243893 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a01ce20-c1b5-4208-bb2a-1fbbde607700" path="/var/lib/kubelet/pods/0a01ce20-c1b5-4208-bb2a-1fbbde607700/volumes" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.244524 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8044d012-63a7-4d09-9b28-9d45c6a49367" path="/var/lib/kubelet/pods/8044d012-63a7-4d09-9b28-9d45c6a49367/volumes" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.245498 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2579a67-a700-4de8-a1d7-0b7a0f45549d" path="/var/lib/kubelet/pods/a2579a67-a700-4de8-a1d7-0b7a0f45549d/volumes" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.281634 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-log-httpd\") pod \"313e22c1-d5ed-485b-a22d-9d76b36c060e\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.281821 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-config-data\") pod \"313e22c1-d5ed-485b-a22d-9d76b36c060e\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.281856 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-combined-ca-bundle\") pod \"313e22c1-d5ed-485b-a22d-9d76b36c060e\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.282361 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-scripts\") pod \"313e22c1-d5ed-485b-a22d-9d76b36c060e\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.282432 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-sg-core-conf-yaml\") pod \"313e22c1-d5ed-485b-a22d-9d76b36c060e\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.282459 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrzv7\" (UniqueName: \"kubernetes.io/projected/313e22c1-d5ed-485b-a22d-9d76b36c060e-kube-api-access-zrzv7\") pod \"313e22c1-d5ed-485b-a22d-9d76b36c060e\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.282493 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-run-httpd\") pod \"313e22c1-d5ed-485b-a22d-9d76b36c060e\" (UID: \"313e22c1-d5ed-485b-a22d-9d76b36c060e\") " Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.283221 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "313e22c1-d5ed-485b-a22d-9d76b36c060e" (UID: "313e22c1-d5ed-485b-a22d-9d76b36c060e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.283987 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "313e22c1-d5ed-485b-a22d-9d76b36c060e" (UID: "313e22c1-d5ed-485b-a22d-9d76b36c060e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.286632 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-scripts" (OuterVolumeSpecName: "scripts") pod "313e22c1-d5ed-485b-a22d-9d76b36c060e" (UID: "313e22c1-d5ed-485b-a22d-9d76b36c060e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.287694 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/313e22c1-d5ed-485b-a22d-9d76b36c060e-kube-api-access-zrzv7" (OuterVolumeSpecName: "kube-api-access-zrzv7") pod "313e22c1-d5ed-485b-a22d-9d76b36c060e" (UID: "313e22c1-d5ed-485b-a22d-9d76b36c060e"). InnerVolumeSpecName "kube-api-access-zrzv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.309701 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "313e22c1-d5ed-485b-a22d-9d76b36c060e" (UID: "313e22c1-d5ed-485b-a22d-9d76b36c060e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.365378 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "313e22c1-d5ed-485b-a22d-9d76b36c060e" (UID: "313e22c1-d5ed-485b-a22d-9d76b36c060e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.384418 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.384456 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrzv7\" (UniqueName: \"kubernetes.io/projected/313e22c1-d5ed-485b-a22d-9d76b36c060e-kube-api-access-zrzv7\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.384466 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.384474 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/313e22c1-d5ed-485b-a22d-9d76b36c060e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.384482 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.384490 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.421577 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-config-data" (OuterVolumeSpecName: "config-data") pod "313e22c1-d5ed-485b-a22d-9d76b36c060e" (UID: "313e22c1-d5ed-485b-a22d-9d76b36c060e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:32 crc kubenswrapper[4788]: I1010 15:06:32.486608 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/313e22c1-d5ed-485b-a22d-9d76b36c060e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.040255 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898","Type":"ContainerStarted","Data":"82415a1ab836383b16f8eb7bf4bf4292b585bf668709c236d3369c260e68828a"} Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.040892 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898","Type":"ContainerStarted","Data":"c55c29e788db344006016d530e0a778ed3eb9c13a119d66a5ef2f5b70e7705af"} Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.043101 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"313e22c1-d5ed-485b-a22d-9d76b36c060e","Type":"ContainerDied","Data":"ba0c63908bd6e05784cc0396573a2ebcd04ae40bbe258cc295f3e7174c8753fa"} Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.043151 4788 scope.go:117] "RemoveContainer" containerID="fb75102a14192d4f5364f015c3d44a361468374ef6a0e3277df9f9bab699db00" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.043196 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.073074 4788 scope.go:117] "RemoveContainer" containerID="09298c4c2ba5002454492fc39933dfc6adedae8c83aa2a613ea9f1d4ec67c96d" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.087369 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.105362 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.120563 4788 scope.go:117] "RemoveContainer" containerID="c3a92a869ed5908e17e2d01bba28698ca41954f1fc612d3ee7e536963a8ff050" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.124372 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:33 crc kubenswrapper[4788]: E1010 15:06:33.124950 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="sg-core" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.124967 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="sg-core" Oct 10 15:06:33 crc kubenswrapper[4788]: E1010 15:06:33.124991 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="ceilometer-central-agent" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.124997 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="ceilometer-central-agent" Oct 10 15:06:33 crc kubenswrapper[4788]: E1010 15:06:33.125007 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="ceilometer-notification-agent" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.125014 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="ceilometer-notification-agent" Oct 10 15:06:33 crc kubenswrapper[4788]: E1010 15:06:33.125027 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="proxy-httpd" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.125033 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="proxy-httpd" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.125229 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="proxy-httpd" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.125243 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="sg-core" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.125255 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="ceilometer-notification-agent" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.125269 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" containerName="ceilometer-central-agent" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.127760 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.130796 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.131227 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.138048 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.195167 4788 scope.go:117] "RemoveContainer" containerID="67cef068036c5b37dba000bedcb9bf0ff6c33de9d830345baad07b8a59071afa" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.308829 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.308898 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-config-data\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.308924 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-run-httpd\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.308964 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.309345 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-scripts\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.309414 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-log-httpd\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.309657 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rxc8\" (UniqueName: \"kubernetes.io/projected/24d495c8-6c56-41b2-b80a-9e6474363c19-kube-api-access-6rxc8\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.411872 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.411956 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-config-data\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.411981 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-run-httpd\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.412021 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.412082 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-scripts\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.412105 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-log-httpd\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.412194 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rxc8\" (UniqueName: \"kubernetes.io/projected/24d495c8-6c56-41b2-b80a-9e6474363c19-kube-api-access-6rxc8\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.412965 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-run-httpd\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.414001 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-log-httpd\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.419658 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.420567 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.429828 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-scripts\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.430745 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-config-data\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.432749 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rxc8\" (UniqueName: \"kubernetes.io/projected/24d495c8-6c56-41b2-b80a-9e6474363c19-kube-api-access-6rxc8\") pod \"ceilometer-0\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.455131 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:33 crc kubenswrapper[4788]: I1010 15:06:33.963412 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:33 crc kubenswrapper[4788]: W1010 15:06:33.970278 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24d495c8_6c56_41b2_b80a_9e6474363c19.slice/crio-526029f5537a6ccb878b01871030224f2b5cdf7efb7ecd1ad6f1c73cf10e43b9 WatchSource:0}: Error finding container 526029f5537a6ccb878b01871030224f2b5cdf7efb7ecd1ad6f1c73cf10e43b9: Status 404 returned error can't find the container with id 526029f5537a6ccb878b01871030224f2b5cdf7efb7ecd1ad6f1c73cf10e43b9 Oct 10 15:06:34 crc kubenswrapper[4788]: I1010 15:06:34.054343 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerStarted","Data":"526029f5537a6ccb878b01871030224f2b5cdf7efb7ecd1ad6f1c73cf10e43b9"} Oct 10 15:06:34 crc kubenswrapper[4788]: I1010 15:06:34.056924 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898","Type":"ContainerStarted","Data":"737b4bebc0138dcddcb915236eb4ef30c70a15399b522e7c7159d53fa58d0ee4"} Oct 10 15:06:34 crc kubenswrapper[4788]: I1010 15:06:34.247696 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="313e22c1-d5ed-485b-a22d-9d76b36c060e" path="/var/lib/kubelet/pods/313e22c1-d5ed-485b-a22d-9d76b36c060e/volumes" Oct 10 15:06:34 crc kubenswrapper[4788]: I1010 15:06:34.895419 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.895403876 podStartE2EDuration="3.895403876s" podCreationTimestamp="2025-10-10 15:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:34.078605275 +0000 UTC m=+1296.528320823" watchObservedRunningTime="2025-10-10 15:06:34.895403876 +0000 UTC m=+1297.345119424" Oct 10 15:06:34 crc kubenswrapper[4788]: I1010 15:06:34.900947 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:35 crc kubenswrapper[4788]: I1010 15:06:35.069644 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerStarted","Data":"c74559860631b8471c48eefeb8cff5db5a27cbd9c356e3b6d62ee684f03c0f98"} Oct 10 15:06:36 crc kubenswrapper[4788]: I1010 15:06:36.084607 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerStarted","Data":"738718baa0a2bcb4b08f63698f8aa53d2a8d14acac079c8121fb629288d77d15"} Oct 10 15:06:36 crc kubenswrapper[4788]: I1010 15:06:36.085405 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerStarted","Data":"75754fabec2a957f5473c8b6d0e269b61de4bd802556fd25db9f3d0c62cdf27f"} Oct 10 15:06:36 crc kubenswrapper[4788]: I1010 15:06:36.514847 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 10 15:06:36 crc kubenswrapper[4788]: I1010 15:06:36.944188 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-wrbdn"] Oct 10 15:06:36 crc kubenswrapper[4788]: I1010 15:06:36.946256 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wrbdn" Oct 10 15:06:36 crc kubenswrapper[4788]: I1010 15:06:36.954039 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wrbdn"] Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.006541 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzt8p\" (UniqueName: \"kubernetes.io/projected/35c81559-a473-4667-8e31-71a388aa4cdf-kube-api-access-hzt8p\") pod \"nova-api-db-create-wrbdn\" (UID: \"35c81559-a473-4667-8e31-71a388aa4cdf\") " pod="openstack/nova-api-db-create-wrbdn" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.055836 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-w2642"] Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.058171 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w2642" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.071521 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-w2642"] Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.108591 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzt8p\" (UniqueName: \"kubernetes.io/projected/35c81559-a473-4667-8e31-71a388aa4cdf-kube-api-access-hzt8p\") pod \"nova-api-db-create-wrbdn\" (UID: \"35c81559-a473-4667-8e31-71a388aa4cdf\") " pod="openstack/nova-api-db-create-wrbdn" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.109368 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4drt\" (UniqueName: \"kubernetes.io/projected/3ab682b0-0d95-4150-8e87-88ad9a7d95a7-kube-api-access-w4drt\") pod \"nova-cell0-db-create-w2642\" (UID: \"3ab682b0-0d95-4150-8e87-88ad9a7d95a7\") " pod="openstack/nova-cell0-db-create-w2642" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.137629 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzt8p\" (UniqueName: \"kubernetes.io/projected/35c81559-a473-4667-8e31-71a388aa4cdf-kube-api-access-hzt8p\") pod \"nova-api-db-create-wrbdn\" (UID: \"35c81559-a473-4667-8e31-71a388aa4cdf\") " pod="openstack/nova-api-db-create-wrbdn" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.211881 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4drt\" (UniqueName: \"kubernetes.io/projected/3ab682b0-0d95-4150-8e87-88ad9a7d95a7-kube-api-access-w4drt\") pod \"nova-cell0-db-create-w2642\" (UID: \"3ab682b0-0d95-4150-8e87-88ad9a7d95a7\") " pod="openstack/nova-cell0-db-create-w2642" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.236669 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4drt\" (UniqueName: \"kubernetes.io/projected/3ab682b0-0d95-4150-8e87-88ad9a7d95a7-kube-api-access-w4drt\") pod \"nova-cell0-db-create-w2642\" (UID: \"3ab682b0-0d95-4150-8e87-88ad9a7d95a7\") " pod="openstack/nova-cell0-db-create-w2642" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.254211 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dlwh4"] Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.255632 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dlwh4" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.263305 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wrbdn" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.286286 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w2642" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.288274 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dlwh4"] Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.420259 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfd8p\" (UniqueName: \"kubernetes.io/projected/fd2cb897-fef8-4152-aacf-409fa2b13209-kube-api-access-wfd8p\") pod \"nova-cell1-db-create-dlwh4\" (UID: \"fd2cb897-fef8-4152-aacf-409fa2b13209\") " pod="openstack/nova-cell1-db-create-dlwh4" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.522122 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfd8p\" (UniqueName: \"kubernetes.io/projected/fd2cb897-fef8-4152-aacf-409fa2b13209-kube-api-access-wfd8p\") pod \"nova-cell1-db-create-dlwh4\" (UID: \"fd2cb897-fef8-4152-aacf-409fa2b13209\") " pod="openstack/nova-cell1-db-create-dlwh4" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.543745 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfd8p\" (UniqueName: \"kubernetes.io/projected/fd2cb897-fef8-4152-aacf-409fa2b13209-kube-api-access-wfd8p\") pod \"nova-cell1-db-create-dlwh4\" (UID: \"fd2cb897-fef8-4152-aacf-409fa2b13209\") " pod="openstack/nova-cell1-db-create-dlwh4" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.624734 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dlwh4" Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.827037 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wrbdn"] Oct 10 15:06:37 crc kubenswrapper[4788]: W1010 15:06:37.833250 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35c81559_a473_4667_8e31_71a388aa4cdf.slice/crio-9c723d3eb9758624ddc850281135680f1c4d81f102425fbd075bb503596ff6e0 WatchSource:0}: Error finding container 9c723d3eb9758624ddc850281135680f1c4d81f102425fbd075bb503596ff6e0: Status 404 returned error can't find the container with id 9c723d3eb9758624ddc850281135680f1c4d81f102425fbd075bb503596ff6e0 Oct 10 15:06:37 crc kubenswrapper[4788]: I1010 15:06:37.923132 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-w2642"] Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.113998 4788 generic.go:334] "Generic (PLEG): container finished" podID="35c81559-a473-4667-8e31-71a388aa4cdf" containerID="6129639cc6300aaa58f2036c176db47627bdf756435da9fb378d8d6f03f3906f" exitCode=0 Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.114081 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wrbdn" event={"ID":"35c81559-a473-4667-8e31-71a388aa4cdf","Type":"ContainerDied","Data":"6129639cc6300aaa58f2036c176db47627bdf756435da9fb378d8d6f03f3906f"} Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.114118 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wrbdn" event={"ID":"35c81559-a473-4667-8e31-71a388aa4cdf","Type":"ContainerStarted","Data":"9c723d3eb9758624ddc850281135680f1c4d81f102425fbd075bb503596ff6e0"} Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.116002 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w2642" event={"ID":"3ab682b0-0d95-4150-8e87-88ad9a7d95a7","Type":"ContainerStarted","Data":"e350481feed6ba25ffcca226950cb056b9cf131136f51f3982b580529d1ff362"} Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.118933 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerStarted","Data":"677ff4519271b2895389af925e7a410ae00b4a773ad88355cb333084a7d66722"} Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.119228 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="ceilometer-central-agent" containerID="cri-o://c74559860631b8471c48eefeb8cff5db5a27cbd9c356e3b6d62ee684f03c0f98" gracePeriod=30 Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.119352 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="ceilometer-notification-agent" containerID="cri-o://75754fabec2a957f5473c8b6d0e269b61de4bd802556fd25db9f3d0c62cdf27f" gracePeriod=30 Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.119388 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="proxy-httpd" containerID="cri-o://677ff4519271b2895389af925e7a410ae00b4a773ad88355cb333084a7d66722" gracePeriod=30 Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.119408 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.119388 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="sg-core" containerID="cri-o://738718baa0a2bcb4b08f63698f8aa53d2a8d14acac079c8121fb629288d77d15" gracePeriod=30 Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.173314 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dlwh4"] Oct 10 15:06:38 crc kubenswrapper[4788]: W1010 15:06:38.181747 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd2cb897_fef8_4152_aacf_409fa2b13209.slice/crio-b584c4cfd4f29b6e1ea8fad04708b1bc10e610934f8123947bb5cc6f4bf13684 WatchSource:0}: Error finding container b584c4cfd4f29b6e1ea8fad04708b1bc10e610934f8123947bb5cc6f4bf13684: Status 404 returned error can't find the container with id b584c4cfd4f29b6e1ea8fad04708b1bc10e610934f8123947bb5cc6f4bf13684 Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.190029 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.154246506 podStartE2EDuration="5.190000964s" podCreationTimestamp="2025-10-10 15:06:33 +0000 UTC" firstStartedPulling="2025-10-10 15:06:33.973298679 +0000 UTC m=+1296.423014227" lastFinishedPulling="2025-10-10 15:06:37.009053137 +0000 UTC m=+1299.458768685" observedRunningTime="2025-10-10 15:06:38.153854028 +0000 UTC m=+1300.603569576" watchObservedRunningTime="2025-10-10 15:06:38.190000964 +0000 UTC m=+1300.639716502" Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.469029 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.470647 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerName="glance-httpd" containerID="cri-o://183b4d966d60fe3ff4c7e99782bb6afa87a35b173ac80cac610a4890b558dd1c" gracePeriod=30 Oct 10 15:06:38 crc kubenswrapper[4788]: I1010 15:06:38.472244 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerName="glance-log" containerID="cri-o://e09f1669e3a187300831d30e49a102971bd1cc7640e3796649e5f7fdfbec27e0" gracePeriod=30 Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.129333 4788 generic.go:334] "Generic (PLEG): container finished" podID="fd2cb897-fef8-4152-aacf-409fa2b13209" containerID="99ac1c7166ca7cb51003167ee8e34ec922f167234c00c272405405fcad496e3a" exitCode=0 Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.129706 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dlwh4" event={"ID":"fd2cb897-fef8-4152-aacf-409fa2b13209","Type":"ContainerDied","Data":"99ac1c7166ca7cb51003167ee8e34ec922f167234c00c272405405fcad496e3a"} Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.129731 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dlwh4" event={"ID":"fd2cb897-fef8-4152-aacf-409fa2b13209","Type":"ContainerStarted","Data":"b584c4cfd4f29b6e1ea8fad04708b1bc10e610934f8123947bb5cc6f4bf13684"} Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.132903 4788 generic.go:334] "Generic (PLEG): container finished" podID="3ab682b0-0d95-4150-8e87-88ad9a7d95a7" containerID="716fea2b9b562081878c7543dd14ca1d4276b93e4640be2eb106923b236a4f59" exitCode=0 Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.132968 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w2642" event={"ID":"3ab682b0-0d95-4150-8e87-88ad9a7d95a7","Type":"ContainerDied","Data":"716fea2b9b562081878c7543dd14ca1d4276b93e4640be2eb106923b236a4f59"} Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.137188 4788 generic.go:334] "Generic (PLEG): container finished" podID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerID="e09f1669e3a187300831d30e49a102971bd1cc7640e3796649e5f7fdfbec27e0" exitCode=143 Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.137244 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc0199c4-f6d9-44e6-a78e-e651c09d0447","Type":"ContainerDied","Data":"e09f1669e3a187300831d30e49a102971bd1cc7640e3796649e5f7fdfbec27e0"} Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.140607 4788 generic.go:334] "Generic (PLEG): container finished" podID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerID="677ff4519271b2895389af925e7a410ae00b4a773ad88355cb333084a7d66722" exitCode=0 Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.140633 4788 generic.go:334] "Generic (PLEG): container finished" podID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerID="738718baa0a2bcb4b08f63698f8aa53d2a8d14acac079c8121fb629288d77d15" exitCode=2 Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.140644 4788 generic.go:334] "Generic (PLEG): container finished" podID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerID="75754fabec2a957f5473c8b6d0e269b61de4bd802556fd25db9f3d0c62cdf27f" exitCode=0 Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.140775 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerDied","Data":"677ff4519271b2895389af925e7a410ae00b4a773ad88355cb333084a7d66722"} Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.140796 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerDied","Data":"738718baa0a2bcb4b08f63698f8aa53d2a8d14acac079c8121fb629288d77d15"} Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.140808 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerDied","Data":"75754fabec2a957f5473c8b6d0e269b61de4bd802556fd25db9f3d0c62cdf27f"} Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.528076 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wrbdn" Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.670886 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzt8p\" (UniqueName: \"kubernetes.io/projected/35c81559-a473-4667-8e31-71a388aa4cdf-kube-api-access-hzt8p\") pod \"35c81559-a473-4667-8e31-71a388aa4cdf\" (UID: \"35c81559-a473-4667-8e31-71a388aa4cdf\") " Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.679830 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35c81559-a473-4667-8e31-71a388aa4cdf-kube-api-access-hzt8p" (OuterVolumeSpecName: "kube-api-access-hzt8p") pod "35c81559-a473-4667-8e31-71a388aa4cdf" (UID: "35c81559-a473-4667-8e31-71a388aa4cdf"). InnerVolumeSpecName "kube-api-access-hzt8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.774016 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzt8p\" (UniqueName: \"kubernetes.io/projected/35c81559-a473-4667-8e31-71a388aa4cdf-kube-api-access-hzt8p\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.908699 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.909347 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerName="glance-log" containerID="cri-o://48c30f8dd20ca989162fe7a6889da2090b8f3f2925b51f181010597cfe4ab535" gracePeriod=30 Oct 10 15:06:39 crc kubenswrapper[4788]: I1010 15:06:39.909460 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerName="glance-httpd" containerID="cri-o://26dc1f755c12a846444239eb5acf3a8b889a93f2aaa3ed202d6dab1c802f9dae" gracePeriod=30 Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.157507 4788 generic.go:334] "Generic (PLEG): container finished" podID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerID="48c30f8dd20ca989162fe7a6889da2090b8f3f2925b51f181010597cfe4ab535" exitCode=143 Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.157588 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9384e19a-7c77-49ca-9dc8-7be525cb4e6c","Type":"ContainerDied","Data":"48c30f8dd20ca989162fe7a6889da2090b8f3f2925b51f181010597cfe4ab535"} Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.159470 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wrbdn" event={"ID":"35c81559-a473-4667-8e31-71a388aa4cdf","Type":"ContainerDied","Data":"9c723d3eb9758624ddc850281135680f1c4d81f102425fbd075bb503596ff6e0"} Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.159501 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c723d3eb9758624ddc850281135680f1c4d81f102425fbd075bb503596ff6e0" Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.159638 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wrbdn" Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.824456 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w2642" Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.839791 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dlwh4" Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.902888 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4drt\" (UniqueName: \"kubernetes.io/projected/3ab682b0-0d95-4150-8e87-88ad9a7d95a7-kube-api-access-w4drt\") pod \"3ab682b0-0d95-4150-8e87-88ad9a7d95a7\" (UID: \"3ab682b0-0d95-4150-8e87-88ad9a7d95a7\") " Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.903036 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfd8p\" (UniqueName: \"kubernetes.io/projected/fd2cb897-fef8-4152-aacf-409fa2b13209-kube-api-access-wfd8p\") pod \"fd2cb897-fef8-4152-aacf-409fa2b13209\" (UID: \"fd2cb897-fef8-4152-aacf-409fa2b13209\") " Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.910967 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab682b0-0d95-4150-8e87-88ad9a7d95a7-kube-api-access-w4drt" (OuterVolumeSpecName: "kube-api-access-w4drt") pod "3ab682b0-0d95-4150-8e87-88ad9a7d95a7" (UID: "3ab682b0-0d95-4150-8e87-88ad9a7d95a7"). InnerVolumeSpecName "kube-api-access-w4drt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:40 crc kubenswrapper[4788]: I1010 15:06:40.912925 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd2cb897-fef8-4152-aacf-409fa2b13209-kube-api-access-wfd8p" (OuterVolumeSpecName: "kube-api-access-wfd8p") pod "fd2cb897-fef8-4152-aacf-409fa2b13209" (UID: "fd2cb897-fef8-4152-aacf-409fa2b13209"). InnerVolumeSpecName "kube-api-access-wfd8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.005902 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4drt\" (UniqueName: \"kubernetes.io/projected/3ab682b0-0d95-4150-8e87-88ad9a7d95a7-kube-api-access-w4drt\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.005948 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfd8p\" (UniqueName: \"kubernetes.io/projected/fd2cb897-fef8-4152-aacf-409fa2b13209-kube-api-access-wfd8p\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.171122 4788 generic.go:334] "Generic (PLEG): container finished" podID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerID="c74559860631b8471c48eefeb8cff5db5a27cbd9c356e3b6d62ee684f03c0f98" exitCode=0 Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.171187 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerDied","Data":"c74559860631b8471c48eefeb8cff5db5a27cbd9c356e3b6d62ee684f03c0f98"} Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.171292 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"24d495c8-6c56-41b2-b80a-9e6474363c19","Type":"ContainerDied","Data":"526029f5537a6ccb878b01871030224f2b5cdf7efb7ecd1ad6f1c73cf10e43b9"} Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.171315 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="526029f5537a6ccb878b01871030224f2b5cdf7efb7ecd1ad6f1c73cf10e43b9" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.173372 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dlwh4" event={"ID":"fd2cb897-fef8-4152-aacf-409fa2b13209","Type":"ContainerDied","Data":"b584c4cfd4f29b6e1ea8fad04708b1bc10e610934f8123947bb5cc6f4bf13684"} Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.173400 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dlwh4" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.173405 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b584c4cfd4f29b6e1ea8fad04708b1bc10e610934f8123947bb5cc6f4bf13684" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.175802 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-w2642" event={"ID":"3ab682b0-0d95-4150-8e87-88ad9a7d95a7","Type":"ContainerDied","Data":"e350481feed6ba25ffcca226950cb056b9cf131136f51f3982b580529d1ff362"} Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.175849 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e350481feed6ba25ffcca226950cb056b9cf131136f51f3982b580529d1ff362" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.175914 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-w2642" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.202723 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.311198 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-sg-core-conf-yaml\") pod \"24d495c8-6c56-41b2-b80a-9e6474363c19\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.311348 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-scripts\") pod \"24d495c8-6c56-41b2-b80a-9e6474363c19\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.311427 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-combined-ca-bundle\") pod \"24d495c8-6c56-41b2-b80a-9e6474363c19\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.311455 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rxc8\" (UniqueName: \"kubernetes.io/projected/24d495c8-6c56-41b2-b80a-9e6474363c19-kube-api-access-6rxc8\") pod \"24d495c8-6c56-41b2-b80a-9e6474363c19\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.311480 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-config-data\") pod \"24d495c8-6c56-41b2-b80a-9e6474363c19\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.311597 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-run-httpd\") pod \"24d495c8-6c56-41b2-b80a-9e6474363c19\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.311678 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-log-httpd\") pod \"24d495c8-6c56-41b2-b80a-9e6474363c19\" (UID: \"24d495c8-6c56-41b2-b80a-9e6474363c19\") " Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.312686 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "24d495c8-6c56-41b2-b80a-9e6474363c19" (UID: "24d495c8-6c56-41b2-b80a-9e6474363c19"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.313708 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "24d495c8-6c56-41b2-b80a-9e6474363c19" (UID: "24d495c8-6c56-41b2-b80a-9e6474363c19"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.325576 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-scripts" (OuterVolumeSpecName: "scripts") pod "24d495c8-6c56-41b2-b80a-9e6474363c19" (UID: "24d495c8-6c56-41b2-b80a-9e6474363c19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.341354 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d495c8-6c56-41b2-b80a-9e6474363c19-kube-api-access-6rxc8" (OuterVolumeSpecName: "kube-api-access-6rxc8") pod "24d495c8-6c56-41b2-b80a-9e6474363c19" (UID: "24d495c8-6c56-41b2-b80a-9e6474363c19"). InnerVolumeSpecName "kube-api-access-6rxc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.348204 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "24d495c8-6c56-41b2-b80a-9e6474363c19" (UID: "24d495c8-6c56-41b2-b80a-9e6474363c19"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.404194 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24d495c8-6c56-41b2-b80a-9e6474363c19" (UID: "24d495c8-6c56-41b2-b80a-9e6474363c19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.413462 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.413497 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.413507 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.413515 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.413525 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rxc8\" (UniqueName: \"kubernetes.io/projected/24d495c8-6c56-41b2-b80a-9e6474363c19-kube-api-access-6rxc8\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.413535 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/24d495c8-6c56-41b2-b80a-9e6474363c19-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.437339 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-config-data" (OuterVolumeSpecName: "config-data") pod "24d495c8-6c56-41b2-b80a-9e6474363c19" (UID: "24d495c8-6c56-41b2-b80a-9e6474363c19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.515206 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24d495c8-6c56-41b2-b80a-9e6474363c19-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:41 crc kubenswrapper[4788]: I1010 15:06:41.753852 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.192822 4788 generic.go:334] "Generic (PLEG): container finished" podID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerID="183b4d966d60fe3ff4c7e99782bb6afa87a35b173ac80cac610a4890b558dd1c" exitCode=0 Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.192906 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc0199c4-f6d9-44e6-a78e-e651c09d0447","Type":"ContainerDied","Data":"183b4d966d60fe3ff4c7e99782bb6afa87a35b173ac80cac610a4890b558dd1c"} Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.193217 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dc0199c4-f6d9-44e6-a78e-e651c09d0447","Type":"ContainerDied","Data":"cac065e6108d17f60762aed680df0db20232700bb1d2a11f245629d395933d2a"} Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.193232 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cac065e6108d17f60762aed680df0db20232700bb1d2a11f245629d395933d2a" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.193259 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.234506 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.258358 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.283920 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.294045 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.294897 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2cb897-fef8-4152-aacf-409fa2b13209" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.294999 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2cb897-fef8-4152-aacf-409fa2b13209" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.295080 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c81559-a473-4667-8e31-71a388aa4cdf" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.295165 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c81559-a473-4667-8e31-71a388aa4cdf" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.295242 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="sg-core" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.295297 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="sg-core" Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.295381 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ab682b0-0d95-4150-8e87-88ad9a7d95a7" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.295434 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ab682b0-0d95-4150-8e87-88ad9a7d95a7" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.295485 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="ceilometer-central-agent" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.295531 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="ceilometer-central-agent" Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.295609 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerName="glance-httpd" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.295658 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerName="glance-httpd" Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.295721 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="proxy-httpd" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.295770 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="proxy-httpd" Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.295823 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerName="glance-log" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.295871 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerName="glance-log" Oct 10 15:06:42 crc kubenswrapper[4788]: E1010 15:06:42.295923 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="ceilometer-notification-agent" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.295992 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="ceilometer-notification-agent" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296215 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="ceilometer-central-agent" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296288 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd2cb897-fef8-4152-aacf-409fa2b13209" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296343 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="ceilometer-notification-agent" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296401 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerName="glance-httpd" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296452 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" containerName="glance-log" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296509 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="proxy-httpd" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296569 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" containerName="sg-core" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296628 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ab682b0-0d95-4150-8e87-88ad9a7d95a7" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.296682 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c81559-a473-4667-8e31-71a388aa4cdf" containerName="mariadb-database-create" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.298301 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.301371 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.304853 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.333881 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-public-tls-certs\") pod \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.334260 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-logs\") pod \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.334302 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-config-data\") pod \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.334360 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-httpd-run\") pod \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.334439 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vdcm\" (UniqueName: \"kubernetes.io/projected/dc0199c4-f6d9-44e6-a78e-e651c09d0447-kube-api-access-6vdcm\") pod \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.334496 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.334530 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-combined-ca-bundle\") pod \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.334574 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-scripts\") pod \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\" (UID: \"dc0199c4-f6d9-44e6-a78e-e651c09d0447\") " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.337704 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-logs" (OuterVolumeSpecName: "logs") pod "dc0199c4-f6d9-44e6-a78e-e651c09d0447" (UID: "dc0199c4-f6d9-44e6-a78e-e651c09d0447"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.337894 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "dc0199c4-f6d9-44e6-a78e-e651c09d0447" (UID: "dc0199c4-f6d9-44e6-a78e-e651c09d0447"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.341861 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "dc0199c4-f6d9-44e6-a78e-e651c09d0447" (UID: "dc0199c4-f6d9-44e6-a78e-e651c09d0447"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.354197 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc0199c4-f6d9-44e6-a78e-e651c09d0447-kube-api-access-6vdcm" (OuterVolumeSpecName: "kube-api-access-6vdcm") pod "dc0199c4-f6d9-44e6-a78e-e651c09d0447" (UID: "dc0199c4-f6d9-44e6-a78e-e651c09d0447"). InnerVolumeSpecName "kube-api-access-6vdcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.359024 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-scripts" (OuterVolumeSpecName: "scripts") pod "dc0199c4-f6d9-44e6-a78e-e651c09d0447" (UID: "dc0199c4-f6d9-44e6-a78e-e651c09d0447"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.361245 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.386468 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc0199c4-f6d9-44e6-a78e-e651c09d0447" (UID: "dc0199c4-f6d9-44e6-a78e-e651c09d0447"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.414227 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dc0199c4-f6d9-44e6-a78e-e651c09d0447" (UID: "dc0199c4-f6d9-44e6-a78e-e651c09d0447"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.426500 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-config-data" (OuterVolumeSpecName: "config-data") pod "dc0199c4-f6d9-44e6-a78e-e651c09d0447" (UID: "dc0199c4-f6d9-44e6-a78e-e651c09d0447"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.437841 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgzbx\" (UniqueName: \"kubernetes.io/projected/e9801af7-97f4-42c6-bae6-9d617fda37ef-kube-api-access-lgzbx\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.437903 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438036 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-config-data\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438081 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-scripts\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438192 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-log-httpd\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438214 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-run-httpd\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438237 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438302 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438322 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438336 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438348 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dc0199c4-f6d9-44e6-a78e-e651c09d0447-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438362 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vdcm\" (UniqueName: \"kubernetes.io/projected/dc0199c4-f6d9-44e6-a78e-e651c09d0447-kube-api-access-6vdcm\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438390 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438402 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.438415 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc0199c4-f6d9-44e6-a78e-e651c09d0447-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.462177 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.539950 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgzbx\" (UniqueName: \"kubernetes.io/projected/e9801af7-97f4-42c6-bae6-9d617fda37ef-kube-api-access-lgzbx\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.539998 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.540030 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-config-data\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.540062 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-scripts\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.540128 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-log-httpd\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.540160 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-run-httpd\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.540177 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.540261 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.540890 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-log-httpd\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.541017 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-run-httpd\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.544544 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-config-data\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.545331 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.546002 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.547789 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-scripts\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.570642 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgzbx\" (UniqueName: \"kubernetes.io/projected/e9801af7-97f4-42c6-bae6-9d617fda37ef-kube-api-access-lgzbx\") pod \"ceilometer-0\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " pod="openstack/ceilometer-0" Oct 10 15:06:42 crc kubenswrapper[4788]: I1010 15:06:42.633917 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.200694 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.235063 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.250600 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.262603 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.291770 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.294157 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.301035 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.303618 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.304808 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.360774 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.360817 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.360840 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.361083 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-config-data\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.361117 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp6sp\" (UniqueName: \"kubernetes.io/projected/e2166478-a6f0-421b-9ea7-17f60ee9605b-kube-api-access-xp6sp\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.361173 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.361195 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-logs\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.361265 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-scripts\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.462994 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463060 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463080 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463103 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-config-data\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463118 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp6sp\" (UniqueName: \"kubernetes.io/projected/e2166478-a6f0-421b-9ea7-17f60ee9605b-kube-api-access-xp6sp\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463151 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463169 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-logs\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463192 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-scripts\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463469 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463767 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.463837 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-logs\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.470845 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-scripts\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.471587 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-config-data\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.471919 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.473657 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.482366 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp6sp\" (UniqueName: \"kubernetes.io/projected/e2166478-a6f0-421b-9ea7-17f60ee9605b-kube-api-access-xp6sp\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.504209 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " pod="openstack/glance-default-external-api-0" Oct 10 15:06:43 crc kubenswrapper[4788]: I1010 15:06:43.615840 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.214310 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerStarted","Data":"277fcf7715ac1e428c6b166447df3e62080351076d1128c13913338bf200eda2"} Oct 10 15:06:44 crc kubenswrapper[4788]: W1010 15:06:44.214339 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2166478_a6f0_421b_9ea7_17f60ee9605b.slice/crio-3022642694ba4ddf812aab1db3f5d23fe1d48af8e85b4b87a5d823f6c2e80f19 WatchSource:0}: Error finding container 3022642694ba4ddf812aab1db3f5d23fe1d48af8e85b4b87a5d823f6c2e80f19: Status 404 returned error can't find the container with id 3022642694ba4ddf812aab1db3f5d23fe1d48af8e85b4b87a5d823f6c2e80f19 Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.220622 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.225759 4788 generic.go:334] "Generic (PLEG): container finished" podID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerID="26dc1f755c12a846444239eb5acf3a8b889a93f2aaa3ed202d6dab1c802f9dae" exitCode=0 Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.225792 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9384e19a-7c77-49ca-9dc8-7be525cb4e6c","Type":"ContainerDied","Data":"26dc1f755c12a846444239eb5acf3a8b889a93f2aaa3ed202d6dab1c802f9dae"} Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.250382 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d495c8-6c56-41b2-b80a-9e6474363c19" path="/var/lib/kubelet/pods/24d495c8-6c56-41b2-b80a-9e6474363c19/volumes" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.252383 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc0199c4-f6d9-44e6-a78e-e651c09d0447" path="/var/lib/kubelet/pods/dc0199c4-f6d9-44e6-a78e-e651c09d0447/volumes" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.723595 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.792603 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-combined-ca-bundle\") pod \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.792855 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-httpd-run\") pod \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.792932 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jff6\" (UniqueName: \"kubernetes.io/projected/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-kube-api-access-7jff6\") pod \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.792995 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-config-data\") pod \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.793698 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9384e19a-7c77-49ca-9dc8-7be525cb4e6c" (UID: "9384e19a-7c77-49ca-9dc8-7be525cb4e6c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.793905 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-scripts\") pod \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.794615 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.794666 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-internal-tls-certs\") pod \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.794891 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-logs\") pod \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\" (UID: \"9384e19a-7c77-49ca-9dc8-7be525cb4e6c\") " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.795923 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.797452 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-logs" (OuterVolumeSpecName: "logs") pod "9384e19a-7c77-49ca-9dc8-7be525cb4e6c" (UID: "9384e19a-7c77-49ca-9dc8-7be525cb4e6c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.802426 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "9384e19a-7c77-49ca-9dc8-7be525cb4e6c" (UID: "9384e19a-7c77-49ca-9dc8-7be525cb4e6c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.802442 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-kube-api-access-7jff6" (OuterVolumeSpecName: "kube-api-access-7jff6") pod "9384e19a-7c77-49ca-9dc8-7be525cb4e6c" (UID: "9384e19a-7c77-49ca-9dc8-7be525cb4e6c"). InnerVolumeSpecName "kube-api-access-7jff6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.802461 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-scripts" (OuterVolumeSpecName: "scripts") pod "9384e19a-7c77-49ca-9dc8-7be525cb4e6c" (UID: "9384e19a-7c77-49ca-9dc8-7be525cb4e6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.858414 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9384e19a-7c77-49ca-9dc8-7be525cb4e6c" (UID: "9384e19a-7c77-49ca-9dc8-7be525cb4e6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.890094 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9384e19a-7c77-49ca-9dc8-7be525cb4e6c" (UID: "9384e19a-7c77-49ca-9dc8-7be525cb4e6c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.899742 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jff6\" (UniqueName: \"kubernetes.io/projected/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-kube-api-access-7jff6\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.899782 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.899811 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.899822 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.899831 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.899839 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.903934 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-config-data" (OuterVolumeSpecName: "config-data") pod "9384e19a-7c77-49ca-9dc8-7be525cb4e6c" (UID: "9384e19a-7c77-49ca-9dc8-7be525cb4e6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:44 crc kubenswrapper[4788]: I1010 15:06:44.933027 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.001730 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9384e19a-7c77-49ca-9dc8-7be525cb4e6c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.001780 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.268395 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerStarted","Data":"01160995ae60c4571cca3b803588097b33fd13ac3422376fa3c5f2388f757f01"} Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.273776 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9384e19a-7c77-49ca-9dc8-7be525cb4e6c","Type":"ContainerDied","Data":"53bbea773567a0f8e1a99ce19cfca345b382643391fb5abf5ff0dbaff2902598"} Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.273855 4788 scope.go:117] "RemoveContainer" containerID="26dc1f755c12a846444239eb5acf3a8b889a93f2aaa3ed202d6dab1c802f9dae" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.273989 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.277382 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2166478-a6f0-421b-9ea7-17f60ee9605b","Type":"ContainerStarted","Data":"b5c371e53f6216ef8edfc3b5203d8b1eccf15ab2af6828b0ddf25a2c86e85ada"} Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.277441 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2166478-a6f0-421b-9ea7-17f60ee9605b","Type":"ContainerStarted","Data":"3022642694ba4ddf812aab1db3f5d23fe1d48af8e85b4b87a5d823f6c2e80f19"} Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.330400 4788 scope.go:117] "RemoveContainer" containerID="48c30f8dd20ca989162fe7a6889da2090b8f3f2925b51f181010597cfe4ab535" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.338389 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.354820 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.387725 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:06:45 crc kubenswrapper[4788]: E1010 15:06:45.388376 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerName="glance-httpd" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.388416 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerName="glance-httpd" Oct 10 15:06:45 crc kubenswrapper[4788]: E1010 15:06:45.388487 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerName="glance-log" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.388495 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerName="glance-log" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.388838 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerName="glance-log" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.388891 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" containerName="glance-httpd" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.392892 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.399978 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.400300 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.419374 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.530352 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.530410 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.530442 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.530471 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-config-data\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.530547 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.530622 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw8hf\" (UniqueName: \"kubernetes.io/projected/df85b9e1-2830-40c3-bbea-555bcddf5545-kube-api-access-lw8hf\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.531013 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-logs\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.531036 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-scripts\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.633121 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.633559 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw8hf\" (UniqueName: \"kubernetes.io/projected/df85b9e1-2830-40c3-bbea-555bcddf5545-kube-api-access-lw8hf\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.633595 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-logs\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.633616 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-scripts\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.633673 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.633703 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.633731 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.633762 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-config-data\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.634431 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-logs\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.634448 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.634826 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.640285 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.641395 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-config-data\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.649719 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-scripts\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.652899 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.655079 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw8hf\" (UniqueName: \"kubernetes.io/projected/df85b9e1-2830-40c3-bbea-555bcddf5545-kube-api-access-lw8hf\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.670854 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " pod="openstack/glance-default-internal-api-0" Oct 10 15:06:45 crc kubenswrapper[4788]: I1010 15:06:45.772481 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:46 crc kubenswrapper[4788]: I1010 15:06:46.249496 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9384e19a-7c77-49ca-9dc8-7be525cb4e6c" path="/var/lib/kubelet/pods/9384e19a-7c77-49ca-9dc8-7be525cb4e6c/volumes" Oct 10 15:06:46 crc kubenswrapper[4788]: I1010 15:06:46.300355 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerStarted","Data":"7850849dbd4da67ba375448c760d35220c8d3f553cf2971c5d018050f1861e83"} Oct 10 15:06:46 crc kubenswrapper[4788]: I1010 15:06:46.300430 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerStarted","Data":"2763b21f8f2c43d99665a3a58ae1bed05c53d708c0729a33aa4a3ea8825a4b18"} Oct 10 15:06:46 crc kubenswrapper[4788]: I1010 15:06:46.308675 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2166478-a6f0-421b-9ea7-17f60ee9605b","Type":"ContainerStarted","Data":"905b0fdae791afc90732669e1fd50c89e576f7fe20ab044f3b24b621f2e22853"} Oct 10 15:06:46 crc kubenswrapper[4788]: I1010 15:06:46.341322 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:06:46 crc kubenswrapper[4788]: W1010 15:06:46.348653 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf85b9e1_2830_40c3_bbea_555bcddf5545.slice/crio-73b7decab65445051dfd83cb6ad434f9bf9dbe77af7ef1c02e9a1752cb60a81b WatchSource:0}: Error finding container 73b7decab65445051dfd83cb6ad434f9bf9dbe77af7ef1c02e9a1752cb60a81b: Status 404 returned error can't find the container with id 73b7decab65445051dfd83cb6ad434f9bf9dbe77af7ef1c02e9a1752cb60a81b Oct 10 15:06:46 crc kubenswrapper[4788]: I1010 15:06:46.358378 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.358359164 podStartE2EDuration="3.358359164s" podCreationTimestamp="2025-10-10 15:06:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:46.332531427 +0000 UTC m=+1308.782246975" watchObservedRunningTime="2025-10-10 15:06:46.358359164 +0000 UTC m=+1308.808074712" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.041427 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-96b1-account-create-vkctd"] Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.043457 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-96b1-account-create-vkctd" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.046004 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.049945 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-96b1-account-create-vkctd"] Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.173739 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9mlr\" (UniqueName: \"kubernetes.io/projected/dc982550-65dc-47de-b336-4e797e813403-kube-api-access-c9mlr\") pod \"nova-api-96b1-account-create-vkctd\" (UID: \"dc982550-65dc-47de-b336-4e797e813403\") " pod="openstack/nova-api-96b1-account-create-vkctd" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.235697 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-bb0b-account-create-bnxjn"] Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.237559 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bb0b-account-create-bnxjn" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.242698 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.246964 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-bb0b-account-create-bnxjn"] Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.276807 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9mlr\" (UniqueName: \"kubernetes.io/projected/dc982550-65dc-47de-b336-4e797e813403-kube-api-access-c9mlr\") pod \"nova-api-96b1-account-create-vkctd\" (UID: \"dc982550-65dc-47de-b336-4e797e813403\") " pod="openstack/nova-api-96b1-account-create-vkctd" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.295420 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9mlr\" (UniqueName: \"kubernetes.io/projected/dc982550-65dc-47de-b336-4e797e813403-kube-api-access-c9mlr\") pod \"nova-api-96b1-account-create-vkctd\" (UID: \"dc982550-65dc-47de-b336-4e797e813403\") " pod="openstack/nova-api-96b1-account-create-vkctd" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.324015 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"df85b9e1-2830-40c3-bbea-555bcddf5545","Type":"ContainerStarted","Data":"ec39cee08096b21f5fb188e2f4be5f41b034a3a3eb4ccdb2d9860218975f6b3f"} Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.324058 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"df85b9e1-2830-40c3-bbea-555bcddf5545","Type":"ContainerStarted","Data":"73b7decab65445051dfd83cb6ad434f9bf9dbe77af7ef1c02e9a1752cb60a81b"} Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.379563 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxgzz\" (UniqueName: \"kubernetes.io/projected/5ecb15a9-ac38-4eec-93f2-2069e6257ac0-kube-api-access-gxgzz\") pod \"nova-cell0-bb0b-account-create-bnxjn\" (UID: \"5ecb15a9-ac38-4eec-93f2-2069e6257ac0\") " pod="openstack/nova-cell0-bb0b-account-create-bnxjn" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.426594 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-96b1-account-create-vkctd" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.439039 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-0a83-account-create-8k2jh"] Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.440467 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0a83-account-create-8k2jh" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.443180 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.467128 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0a83-account-create-8k2jh"] Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.486480 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz4f5\" (UniqueName: \"kubernetes.io/projected/244bd917-45fe-43e1-8386-029d12f5cb9a-kube-api-access-dz4f5\") pod \"nova-cell1-0a83-account-create-8k2jh\" (UID: \"244bd917-45fe-43e1-8386-029d12f5cb9a\") " pod="openstack/nova-cell1-0a83-account-create-8k2jh" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.486572 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxgzz\" (UniqueName: \"kubernetes.io/projected/5ecb15a9-ac38-4eec-93f2-2069e6257ac0-kube-api-access-gxgzz\") pod \"nova-cell0-bb0b-account-create-bnxjn\" (UID: \"5ecb15a9-ac38-4eec-93f2-2069e6257ac0\") " pod="openstack/nova-cell0-bb0b-account-create-bnxjn" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.512361 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxgzz\" (UniqueName: \"kubernetes.io/projected/5ecb15a9-ac38-4eec-93f2-2069e6257ac0-kube-api-access-gxgzz\") pod \"nova-cell0-bb0b-account-create-bnxjn\" (UID: \"5ecb15a9-ac38-4eec-93f2-2069e6257ac0\") " pod="openstack/nova-cell0-bb0b-account-create-bnxjn" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.589925 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz4f5\" (UniqueName: \"kubernetes.io/projected/244bd917-45fe-43e1-8386-029d12f5cb9a-kube-api-access-dz4f5\") pod \"nova-cell1-0a83-account-create-8k2jh\" (UID: \"244bd917-45fe-43e1-8386-029d12f5cb9a\") " pod="openstack/nova-cell1-0a83-account-create-8k2jh" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.615789 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz4f5\" (UniqueName: \"kubernetes.io/projected/244bd917-45fe-43e1-8386-029d12f5cb9a-kube-api-access-dz4f5\") pod \"nova-cell1-0a83-account-create-8k2jh\" (UID: \"244bd917-45fe-43e1-8386-029d12f5cb9a\") " pod="openstack/nova-cell1-0a83-account-create-8k2jh" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.640212 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bb0b-account-create-bnxjn" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.775131 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0a83-account-create-8k2jh" Oct 10 15:06:47 crc kubenswrapper[4788]: I1010 15:06:47.949691 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-96b1-account-create-vkctd"] Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.276320 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-bb0b-account-create-bnxjn"] Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.341925 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerStarted","Data":"687cd476941c190b5c39bbd41874733c3ba048f47ce804de6317c8b045739271"} Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.343871 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.349197 4788 generic.go:334] "Generic (PLEG): container finished" podID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerID="75bc72d77ca51763e142eda1335a09a768c99c2bf9ade041fdec8f17c2aa8354" exitCode=137 Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.349342 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e7ccf50d-1227-42b5-aaf5-3cb3471faad4","Type":"ContainerDied","Data":"75bc72d77ca51763e142eda1335a09a768c99c2bf9ade041fdec8f17c2aa8354"} Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.349377 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e7ccf50d-1227-42b5-aaf5-3cb3471faad4","Type":"ContainerDied","Data":"3479250bfe5ea8d501ffa37482b5bef834f1df8367d26d3b374377f675c899f5"} Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.349391 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3479250bfe5ea8d501ffa37482b5bef834f1df8367d26d3b374377f675c899f5" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.353546 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bb0b-account-create-bnxjn" event={"ID":"5ecb15a9-ac38-4eec-93f2-2069e6257ac0","Type":"ContainerStarted","Data":"3e637ef7df266534af1c94270b72cb05afbeba3b0f2249f9cfe92b951e70d095"} Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.360434 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-96b1-account-create-vkctd" event={"ID":"dc982550-65dc-47de-b336-4e797e813403","Type":"ContainerStarted","Data":"7e46b1cf1f8c6202fa33e2921e3f1273618b996ccc581604d916fffc10f5504e"} Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.360503 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-96b1-account-create-vkctd" event={"ID":"dc982550-65dc-47de-b336-4e797e813403","Type":"ContainerStarted","Data":"45672d05aaf5d2c1474945f84228c1ad1096061f3e2ec17d4a0fcce34b6c689f"} Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.364967 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"df85b9e1-2830-40c3-bbea-555bcddf5545","Type":"ContainerStarted","Data":"10a6407ac6ed46784d87d17ac1f81a2239476dc6186c82ebe7235c2468e4f18c"} Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.371412 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0a83-account-create-8k2jh"] Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.385912 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.446532605 podStartE2EDuration="6.385883208s" podCreationTimestamp="2025-10-10 15:06:42 +0000 UTC" firstStartedPulling="2025-10-10 15:06:43.242433002 +0000 UTC m=+1305.692148550" lastFinishedPulling="2025-10-10 15:06:47.181783605 +0000 UTC m=+1309.631499153" observedRunningTime="2025-10-10 15:06:48.37116267 +0000 UTC m=+1310.820878218" watchObservedRunningTime="2025-10-10 15:06:48.385883208 +0000 UTC m=+1310.835598756" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.406159 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.406113054 podStartE2EDuration="3.406113054s" podCreationTimestamp="2025-10-10 15:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:48.393058461 +0000 UTC m=+1310.842774009" watchObservedRunningTime="2025-10-10 15:06:48.406113054 +0000 UTC m=+1310.855828592" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.460947 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.527803 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-etc-machine-id\") pod \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.528105 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59jjz\" (UniqueName: \"kubernetes.io/projected/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-kube-api-access-59jjz\") pod \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.528243 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data-custom\") pod \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.528394 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-scripts\") pod \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.528476 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-logs\") pod \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.528622 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-combined-ca-bundle\") pod \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.528841 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data\") pod \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\" (UID: \"e7ccf50d-1227-42b5-aaf5-3cb3471faad4\") " Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.532326 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e7ccf50d-1227-42b5-aaf5-3cb3471faad4" (UID: "e7ccf50d-1227-42b5-aaf5-3cb3471faad4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.533330 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-logs" (OuterVolumeSpecName: "logs") pod "e7ccf50d-1227-42b5-aaf5-3cb3471faad4" (UID: "e7ccf50d-1227-42b5-aaf5-3cb3471faad4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.537947 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e7ccf50d-1227-42b5-aaf5-3cb3471faad4" (UID: "e7ccf50d-1227-42b5-aaf5-3cb3471faad4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.542279 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-scripts" (OuterVolumeSpecName: "scripts") pod "e7ccf50d-1227-42b5-aaf5-3cb3471faad4" (UID: "e7ccf50d-1227-42b5-aaf5-3cb3471faad4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.543215 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-kube-api-access-59jjz" (OuterVolumeSpecName: "kube-api-access-59jjz") pod "e7ccf50d-1227-42b5-aaf5-3cb3471faad4" (UID: "e7ccf50d-1227-42b5-aaf5-3cb3471faad4"). InnerVolumeSpecName "kube-api-access-59jjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.624587 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7ccf50d-1227-42b5-aaf5-3cb3471faad4" (UID: "e7ccf50d-1227-42b5-aaf5-3cb3471faad4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.631374 4788 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.631414 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59jjz\" (UniqueName: \"kubernetes.io/projected/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-kube-api-access-59jjz\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.631429 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.631440 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.631452 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.631463 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.672879 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data" (OuterVolumeSpecName: "config-data") pod "e7ccf50d-1227-42b5-aaf5-3cb3471faad4" (UID: "e7ccf50d-1227-42b5-aaf5-3cb3471faad4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:06:48 crc kubenswrapper[4788]: I1010 15:06:48.733856 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ccf50d-1227-42b5-aaf5-3cb3471faad4-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.377757 4788 generic.go:334] "Generic (PLEG): container finished" podID="dc982550-65dc-47de-b336-4e797e813403" containerID="7e46b1cf1f8c6202fa33e2921e3f1273618b996ccc581604d916fffc10f5504e" exitCode=0 Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.377854 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-96b1-account-create-vkctd" event={"ID":"dc982550-65dc-47de-b336-4e797e813403","Type":"ContainerDied","Data":"7e46b1cf1f8c6202fa33e2921e3f1273618b996ccc581604d916fffc10f5504e"} Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.383597 4788 generic.go:334] "Generic (PLEG): container finished" podID="244bd917-45fe-43e1-8386-029d12f5cb9a" containerID="af0d26bb8887c19c55d06a25ecfaf257af7796e1125a3c42fd81381267d630df" exitCode=0 Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.383678 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0a83-account-create-8k2jh" event={"ID":"244bd917-45fe-43e1-8386-029d12f5cb9a","Type":"ContainerDied","Data":"af0d26bb8887c19c55d06a25ecfaf257af7796e1125a3c42fd81381267d630df"} Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.383704 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0a83-account-create-8k2jh" event={"ID":"244bd917-45fe-43e1-8386-029d12f5cb9a","Type":"ContainerStarted","Data":"90355282058fa3fbc31c43d3acc31b5619b25eb375f5f71107f0cb6345b86219"} Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.387811 4788 generic.go:334] "Generic (PLEG): container finished" podID="5ecb15a9-ac38-4eec-93f2-2069e6257ac0" containerID="5c1c142fd945aca3169ac20af4021c94a352596999aa62187f338f4da44d457b" exitCode=0 Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.388816 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bb0b-account-create-bnxjn" event={"ID":"5ecb15a9-ac38-4eec-93f2-2069e6257ac0","Type":"ContainerDied","Data":"5c1c142fd945aca3169ac20af4021c94a352596999aa62187f338f4da44d457b"} Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.388872 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.479208 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.521338 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.538918 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:49 crc kubenswrapper[4788]: E1010 15:06:49.539552 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerName="cinder-api" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.539580 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerName="cinder-api" Oct 10 15:06:49 crc kubenswrapper[4788]: E1010 15:06:49.539592 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerName="cinder-api-log" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.539601 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerName="cinder-api-log" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.539869 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerName="cinder-api" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.539896 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" containerName="cinder-api-log" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.542053 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.546370 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.546579 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.548353 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.557906 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.558006 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.558063 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.558104 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.558252 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.558331 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.558415 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-logs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.558577 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6sj2\" (UniqueName: \"kubernetes.io/projected/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-kube-api-access-v6sj2\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.558719 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-scripts\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.572844 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.663897 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.664073 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-logs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.664623 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-logs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.664799 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6sj2\" (UniqueName: \"kubernetes.io/projected/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-kube-api-access-v6sj2\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.664864 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-scripts\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.664920 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.664972 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.665001 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.665029 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.665067 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.668838 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.679360 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.681768 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.686929 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.688857 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-scripts\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.691519 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.711805 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.715763 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6sj2\" (UniqueName: \"kubernetes.io/projected/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-kube-api-access-v6sj2\") pod \"cinder-api-0\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.850890 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-96b1-account-create-vkctd" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.869442 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9mlr\" (UniqueName: \"kubernetes.io/projected/dc982550-65dc-47de-b336-4e797e813403-kube-api-access-c9mlr\") pod \"dc982550-65dc-47de-b336-4e797e813403\" (UID: \"dc982550-65dc-47de-b336-4e797e813403\") " Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.873337 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.895997 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc982550-65dc-47de-b336-4e797e813403-kube-api-access-c9mlr" (OuterVolumeSpecName: "kube-api-access-c9mlr") pod "dc982550-65dc-47de-b336-4e797e813403" (UID: "dc982550-65dc-47de-b336-4e797e813403"). InnerVolumeSpecName "kube-api-access-c9mlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:49 crc kubenswrapper[4788]: I1010 15:06:49.971701 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9mlr\" (UniqueName: \"kubernetes.io/projected/dc982550-65dc-47de-b336-4e797e813403-kube-api-access-c9mlr\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:50 crc kubenswrapper[4788]: W1010 15:06:50.179258 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab6c469a_3a29_4fdb_b0ce_5671db6645d2.slice/crio-a3cb17cdb6ba13e494afad0db5282252dc94d4f28170a0f1d6259e36b5f20949 WatchSource:0}: Error finding container a3cb17cdb6ba13e494afad0db5282252dc94d4f28170a0f1d6259e36b5f20949: Status 404 returned error can't find the container with id a3cb17cdb6ba13e494afad0db5282252dc94d4f28170a0f1d6259e36b5f20949 Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.183229 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.251922 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7ccf50d-1227-42b5-aaf5-3cb3471faad4" path="/var/lib/kubelet/pods/e7ccf50d-1227-42b5-aaf5-3cb3471faad4/volumes" Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.410528 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ab6c469a-3a29-4fdb-b0ce-5671db6645d2","Type":"ContainerStarted","Data":"a3cb17cdb6ba13e494afad0db5282252dc94d4f28170a0f1d6259e36b5f20949"} Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.414161 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-96b1-account-create-vkctd" event={"ID":"dc982550-65dc-47de-b336-4e797e813403","Type":"ContainerDied","Data":"45672d05aaf5d2c1474945f84228c1ad1096061f3e2ec17d4a0fcce34b6c689f"} Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.414210 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45672d05aaf5d2c1474945f84228c1ad1096061f3e2ec17d4a0fcce34b6c689f" Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.414342 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-96b1-account-create-vkctd" Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.758102 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bb0b-account-create-bnxjn" Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.785573 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0a83-account-create-8k2jh" Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.896001 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dz4f5\" (UniqueName: \"kubernetes.io/projected/244bd917-45fe-43e1-8386-029d12f5cb9a-kube-api-access-dz4f5\") pod \"244bd917-45fe-43e1-8386-029d12f5cb9a\" (UID: \"244bd917-45fe-43e1-8386-029d12f5cb9a\") " Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.897033 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxgzz\" (UniqueName: \"kubernetes.io/projected/5ecb15a9-ac38-4eec-93f2-2069e6257ac0-kube-api-access-gxgzz\") pod \"5ecb15a9-ac38-4eec-93f2-2069e6257ac0\" (UID: \"5ecb15a9-ac38-4eec-93f2-2069e6257ac0\") " Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.901609 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/244bd917-45fe-43e1-8386-029d12f5cb9a-kube-api-access-dz4f5" (OuterVolumeSpecName: "kube-api-access-dz4f5") pod "244bd917-45fe-43e1-8386-029d12f5cb9a" (UID: "244bd917-45fe-43e1-8386-029d12f5cb9a"). InnerVolumeSpecName "kube-api-access-dz4f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.903808 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ecb15a9-ac38-4eec-93f2-2069e6257ac0-kube-api-access-gxgzz" (OuterVolumeSpecName: "kube-api-access-gxgzz") pod "5ecb15a9-ac38-4eec-93f2-2069e6257ac0" (UID: "5ecb15a9-ac38-4eec-93f2-2069e6257ac0"). InnerVolumeSpecName "kube-api-access-gxgzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.998770 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dz4f5\" (UniqueName: \"kubernetes.io/projected/244bd917-45fe-43e1-8386-029d12f5cb9a-kube-api-access-dz4f5\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:50 crc kubenswrapper[4788]: I1010 15:06:50.999031 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxgzz\" (UniqueName: \"kubernetes.io/projected/5ecb15a9-ac38-4eec-93f2-2069e6257ac0-kube-api-access-gxgzz\") on node \"crc\" DevicePath \"\"" Oct 10 15:06:51 crc kubenswrapper[4788]: I1010 15:06:51.427388 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ab6c469a-3a29-4fdb-b0ce-5671db6645d2","Type":"ContainerStarted","Data":"a2c0776de729185f9cd35be86586a5c1c49c778ed6a5e56f595dcf61caf342d5"} Oct 10 15:06:51 crc kubenswrapper[4788]: I1010 15:06:51.428853 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0a83-account-create-8k2jh" event={"ID":"244bd917-45fe-43e1-8386-029d12f5cb9a","Type":"ContainerDied","Data":"90355282058fa3fbc31c43d3acc31b5619b25eb375f5f71107f0cb6345b86219"} Oct 10 15:06:51 crc kubenswrapper[4788]: I1010 15:06:51.428882 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0a83-account-create-8k2jh" Oct 10 15:06:51 crc kubenswrapper[4788]: I1010 15:06:51.428886 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90355282058fa3fbc31c43d3acc31b5619b25eb375f5f71107f0cb6345b86219" Oct 10 15:06:51 crc kubenswrapper[4788]: I1010 15:06:51.432667 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bb0b-account-create-bnxjn" event={"ID":"5ecb15a9-ac38-4eec-93f2-2069e6257ac0","Type":"ContainerDied","Data":"3e637ef7df266534af1c94270b72cb05afbeba3b0f2249f9cfe92b951e70d095"} Oct 10 15:06:51 crc kubenswrapper[4788]: I1010 15:06:51.432692 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e637ef7df266534af1c94270b72cb05afbeba3b0f2249f9cfe92b951e70d095" Oct 10 15:06:51 crc kubenswrapper[4788]: I1010 15:06:51.432755 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bb0b-account-create-bnxjn" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.467077 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ab6c469a-3a29-4fdb-b0ce-5671db6645d2","Type":"ContainerStarted","Data":"dd729b335f5887ad2a47d9138d9341e607bccc05df0bd17180c3c60156b9f713"} Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.468572 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.525316 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.525286924 podStartE2EDuration="3.525286924s" podCreationTimestamp="2025-10-10 15:06:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:06:52.515457528 +0000 UTC m=+1314.965173076" watchObservedRunningTime="2025-10-10 15:06:52.525286924 +0000 UTC m=+1314.975002472" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.539357 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d65h8"] Oct 10 15:06:52 crc kubenswrapper[4788]: E1010 15:06:52.539845 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc982550-65dc-47de-b336-4e797e813403" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.539865 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc982550-65dc-47de-b336-4e797e813403" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: E1010 15:06:52.539891 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ecb15a9-ac38-4eec-93f2-2069e6257ac0" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.539898 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ecb15a9-ac38-4eec-93f2-2069e6257ac0" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: E1010 15:06:52.539912 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244bd917-45fe-43e1-8386-029d12f5cb9a" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.539918 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="244bd917-45fe-43e1-8386-029d12f5cb9a" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.540175 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="244bd917-45fe-43e1-8386-029d12f5cb9a" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.540193 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc982550-65dc-47de-b336-4e797e813403" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.540207 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ecb15a9-ac38-4eec-93f2-2069e6257ac0" containerName="mariadb-account-create" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.540968 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.560092 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.561464 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-r6d9t" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.561511 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.633763 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d65h8"] Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.647090 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7vfn\" (UniqueName: \"kubernetes.io/projected/bd483623-d8ec-4daa-9641-a25e3c92b0eb-kube-api-access-k7vfn\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.647313 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.647409 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-config-data\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.647512 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-scripts\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.753558 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-config-data\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.753646 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-scripts\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.753755 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7vfn\" (UniqueName: \"kubernetes.io/projected/bd483623-d8ec-4daa-9641-a25e3c92b0eb-kube-api-access-k7vfn\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.753783 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.763901 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.774709 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-scripts\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.775978 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-config-data\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.800211 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7vfn\" (UniqueName: \"kubernetes.io/projected/bd483623-d8ec-4daa-9641-a25e3c92b0eb-kube-api-access-k7vfn\") pod \"nova-cell0-conductor-db-sync-d65h8\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:52 crc kubenswrapper[4788]: I1010 15:06:52.862542 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:06:53 crc kubenswrapper[4788]: I1010 15:06:53.404864 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d65h8"] Oct 10 15:06:53 crc kubenswrapper[4788]: W1010 15:06:53.413173 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd483623_d8ec_4daa_9641_a25e3c92b0eb.slice/crio-011b50a5a5fce056fcc11d839aca55296d579d0126e3e0292618659150bbff97 WatchSource:0}: Error finding container 011b50a5a5fce056fcc11d839aca55296d579d0126e3e0292618659150bbff97: Status 404 returned error can't find the container with id 011b50a5a5fce056fcc11d839aca55296d579d0126e3e0292618659150bbff97 Oct 10 15:06:53 crc kubenswrapper[4788]: I1010 15:06:53.475797 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-d65h8" event={"ID":"bd483623-d8ec-4daa-9641-a25e3c92b0eb","Type":"ContainerStarted","Data":"011b50a5a5fce056fcc11d839aca55296d579d0126e3e0292618659150bbff97"} Oct 10 15:06:53 crc kubenswrapper[4788]: I1010 15:06:53.615975 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 15:06:53 crc kubenswrapper[4788]: I1010 15:06:53.616024 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 15:06:53 crc kubenswrapper[4788]: I1010 15:06:53.647162 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 15:06:53 crc kubenswrapper[4788]: I1010 15:06:53.656286 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 15:06:54 crc kubenswrapper[4788]: I1010 15:06:54.486389 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 15:06:54 crc kubenswrapper[4788]: I1010 15:06:54.486716 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 15:06:55 crc kubenswrapper[4788]: I1010 15:06:55.773548 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:55 crc kubenswrapper[4788]: I1010 15:06:55.773589 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:55 crc kubenswrapper[4788]: I1010 15:06:55.816614 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:55 crc kubenswrapper[4788]: I1010 15:06:55.842737 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:56 crc kubenswrapper[4788]: I1010 15:06:56.419072 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 15:06:56 crc kubenswrapper[4788]: I1010 15:06:56.443825 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 15:06:56 crc kubenswrapper[4788]: I1010 15:06:56.509487 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:56 crc kubenswrapper[4788]: I1010 15:06:56.509699 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:58 crc kubenswrapper[4788]: I1010 15:06:58.453037 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 15:06:58 crc kubenswrapper[4788]: I1010 15:06:58.478218 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 15:07:01 crc kubenswrapper[4788]: I1010 15:07:01.582069 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-d65h8" event={"ID":"bd483623-d8ec-4daa-9641-a25e3c92b0eb","Type":"ContainerStarted","Data":"1e94fd37cb99c18e9e70ea41a5932cd392fae9e72bcd73f02398b3f4d1d01fca"} Oct 10 15:07:01 crc kubenswrapper[4788]: I1010 15:07:01.618556 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-d65h8" podStartSLOduration=2.189023052 podStartE2EDuration="9.618527676s" podCreationTimestamp="2025-10-10 15:06:52 +0000 UTC" firstStartedPulling="2025-10-10 15:06:53.415248332 +0000 UTC m=+1315.864963880" lastFinishedPulling="2025-10-10 15:07:00.844752946 +0000 UTC m=+1323.294468504" observedRunningTime="2025-10-10 15:07:01.606740218 +0000 UTC m=+1324.056455786" watchObservedRunningTime="2025-10-10 15:07:01.618527676 +0000 UTC m=+1324.068243244" Oct 10 15:07:01 crc kubenswrapper[4788]: I1010 15:07:01.704536 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 10 15:07:04 crc kubenswrapper[4788]: I1010 15:07:04.751943 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:04 crc kubenswrapper[4788]: I1010 15:07:04.753303 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="ceilometer-central-agent" containerID="cri-o://01160995ae60c4571cca3b803588097b33fd13ac3422376fa3c5f2388f757f01" gracePeriod=30 Oct 10 15:07:04 crc kubenswrapper[4788]: I1010 15:07:04.753860 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="sg-core" containerID="cri-o://7850849dbd4da67ba375448c760d35220c8d3f553cf2971c5d018050f1861e83" gracePeriod=30 Oct 10 15:07:04 crc kubenswrapper[4788]: I1010 15:07:04.753870 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="ceilometer-notification-agent" containerID="cri-o://2763b21f8f2c43d99665a3a58ae1bed05c53d708c0729a33aa4a3ea8825a4b18" gracePeriod=30 Oct 10 15:07:04 crc kubenswrapper[4788]: I1010 15:07:04.753895 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="proxy-httpd" containerID="cri-o://687cd476941c190b5c39bbd41874733c3ba048f47ce804de6317c8b045739271" gracePeriod=30 Oct 10 15:07:04 crc kubenswrapper[4788]: I1010 15:07:04.771588 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.173:3000/\": EOF" Oct 10 15:07:05 crc kubenswrapper[4788]: I1010 15:07:05.631584 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerID="687cd476941c190b5c39bbd41874733c3ba048f47ce804de6317c8b045739271" exitCode=0 Oct 10 15:07:05 crc kubenswrapper[4788]: I1010 15:07:05.632009 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerID="7850849dbd4da67ba375448c760d35220c8d3f553cf2971c5d018050f1861e83" exitCode=2 Oct 10 15:07:05 crc kubenswrapper[4788]: I1010 15:07:05.632019 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerID="01160995ae60c4571cca3b803588097b33fd13ac3422376fa3c5f2388f757f01" exitCode=0 Oct 10 15:07:05 crc kubenswrapper[4788]: I1010 15:07:05.631721 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerDied","Data":"687cd476941c190b5c39bbd41874733c3ba048f47ce804de6317c8b045739271"} Oct 10 15:07:05 crc kubenswrapper[4788]: I1010 15:07:05.632063 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerDied","Data":"7850849dbd4da67ba375448c760d35220c8d3f553cf2971c5d018050f1861e83"} Oct 10 15:07:05 crc kubenswrapper[4788]: I1010 15:07:05.632081 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerDied","Data":"01160995ae60c4571cca3b803588097b33fd13ac3422376fa3c5f2388f757f01"} Oct 10 15:07:07 crc kubenswrapper[4788]: I1010 15:07:07.654010 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerID="2763b21f8f2c43d99665a3a58ae1bed05c53d708c0729a33aa4a3ea8825a4b18" exitCode=0 Oct 10 15:07:07 crc kubenswrapper[4788]: I1010 15:07:07.654086 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerDied","Data":"2763b21f8f2c43d99665a3a58ae1bed05c53d708c0729a33aa4a3ea8825a4b18"} Oct 10 15:07:07 crc kubenswrapper[4788]: I1010 15:07:07.924750 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.014689 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-combined-ca-bundle\") pod \"e9801af7-97f4-42c6-bae6-9d617fda37ef\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.014813 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-log-httpd\") pod \"e9801af7-97f4-42c6-bae6-9d617fda37ef\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.014880 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgzbx\" (UniqueName: \"kubernetes.io/projected/e9801af7-97f4-42c6-bae6-9d617fda37ef-kube-api-access-lgzbx\") pod \"e9801af7-97f4-42c6-bae6-9d617fda37ef\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.015047 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-config-data\") pod \"e9801af7-97f4-42c6-bae6-9d617fda37ef\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.015088 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-scripts\") pod \"e9801af7-97f4-42c6-bae6-9d617fda37ef\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.015197 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-run-httpd\") pod \"e9801af7-97f4-42c6-bae6-9d617fda37ef\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.015244 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-sg-core-conf-yaml\") pod \"e9801af7-97f4-42c6-bae6-9d617fda37ef\" (UID: \"e9801af7-97f4-42c6-bae6-9d617fda37ef\") " Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.015741 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e9801af7-97f4-42c6-bae6-9d617fda37ef" (UID: "e9801af7-97f4-42c6-bae6-9d617fda37ef"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.015971 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e9801af7-97f4-42c6-bae6-9d617fda37ef" (UID: "e9801af7-97f4-42c6-bae6-9d617fda37ef"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.016523 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.016554 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9801af7-97f4-42c6-bae6-9d617fda37ef-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.023849 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9801af7-97f4-42c6-bae6-9d617fda37ef-kube-api-access-lgzbx" (OuterVolumeSpecName: "kube-api-access-lgzbx") pod "e9801af7-97f4-42c6-bae6-9d617fda37ef" (UID: "e9801af7-97f4-42c6-bae6-9d617fda37ef"). InnerVolumeSpecName "kube-api-access-lgzbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.024661 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-scripts" (OuterVolumeSpecName: "scripts") pod "e9801af7-97f4-42c6-bae6-9d617fda37ef" (UID: "e9801af7-97f4-42c6-bae6-9d617fda37ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.060106 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e9801af7-97f4-42c6-bae6-9d617fda37ef" (UID: "e9801af7-97f4-42c6-bae6-9d617fda37ef"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.107426 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9801af7-97f4-42c6-bae6-9d617fda37ef" (UID: "e9801af7-97f4-42c6-bae6-9d617fda37ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.118437 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.118472 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgzbx\" (UniqueName: \"kubernetes.io/projected/e9801af7-97f4-42c6-bae6-9d617fda37ef-kube-api-access-lgzbx\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.118483 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.118492 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.144057 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-config-data" (OuterVolumeSpecName: "config-data") pod "e9801af7-97f4-42c6-bae6-9d617fda37ef" (UID: "e9801af7-97f4-42c6-bae6-9d617fda37ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.221304 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9801af7-97f4-42c6-bae6-9d617fda37ef-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.669265 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e9801af7-97f4-42c6-bae6-9d617fda37ef","Type":"ContainerDied","Data":"277fcf7715ac1e428c6b166447df3e62080351076d1128c13913338bf200eda2"} Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.669333 4788 scope.go:117] "RemoveContainer" containerID="687cd476941c190b5c39bbd41874733c3ba048f47ce804de6317c8b045739271" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.669373 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.710494 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.715936 4788 scope.go:117] "RemoveContainer" containerID="7850849dbd4da67ba375448c760d35220c8d3f553cf2971c5d018050f1861e83" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.727846 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.769265 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:08 crc kubenswrapper[4788]: E1010 15:07:08.769749 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="ceilometer-central-agent" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.769765 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="ceilometer-central-agent" Oct 10 15:07:08 crc kubenswrapper[4788]: E1010 15:07:08.769784 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="sg-core" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.769792 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="sg-core" Oct 10 15:07:08 crc kubenswrapper[4788]: E1010 15:07:08.769810 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="proxy-httpd" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.769820 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="proxy-httpd" Oct 10 15:07:08 crc kubenswrapper[4788]: E1010 15:07:08.769846 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="ceilometer-notification-agent" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.769870 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="ceilometer-notification-agent" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.770094 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="ceilometer-central-agent" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.770118 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="ceilometer-notification-agent" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.770159 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="proxy-httpd" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.770185 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" containerName="sg-core" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.775739 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.777596 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.779439 4788 scope.go:117] "RemoveContainer" containerID="2763b21f8f2c43d99665a3a58ae1bed05c53d708c0729a33aa4a3ea8825a4b18" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.781579 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.781818 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.823184 4788 scope.go:117] "RemoveContainer" containerID="01160995ae60c4571cca3b803588097b33fd13ac3422376fa3c5f2388f757f01" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.836240 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48bkc\" (UniqueName: \"kubernetes.io/projected/feb1c7a6-0737-4d42-8c4e-be384b5a4074-kube-api-access-48bkc\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.836307 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-log-httpd\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.836360 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-run-httpd\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.836390 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.836423 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-scripts\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.836446 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.836508 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-config-data\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.937887 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.938181 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-scripts\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.938212 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.938289 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-config-data\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.938367 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48bkc\" (UniqueName: \"kubernetes.io/projected/feb1c7a6-0737-4d42-8c4e-be384b5a4074-kube-api-access-48bkc\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.938387 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-log-httpd\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.938421 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-run-httpd\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.939514 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-log-httpd\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.939802 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-run-httpd\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.943259 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-scripts\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.943259 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.944650 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-config-data\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.952565 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:08 crc kubenswrapper[4788]: I1010 15:07:08.962730 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48bkc\" (UniqueName: \"kubernetes.io/projected/feb1c7a6-0737-4d42-8c4e-be384b5a4074-kube-api-access-48bkc\") pod \"ceilometer-0\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " pod="openstack/ceilometer-0" Oct 10 15:07:09 crc kubenswrapper[4788]: I1010 15:07:09.111577 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:07:09 crc kubenswrapper[4788]: I1010 15:07:09.408603 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:09 crc kubenswrapper[4788]: W1010 15:07:09.414783 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfeb1c7a6_0737_4d42_8c4e_be384b5a4074.slice/crio-a28995f507205de0012a07bbf23fe5bdfc9d314e572c4f976638c6fc1e49da8e WatchSource:0}: Error finding container a28995f507205de0012a07bbf23fe5bdfc9d314e572c4f976638c6fc1e49da8e: Status 404 returned error can't find the container with id a28995f507205de0012a07bbf23fe5bdfc9d314e572c4f976638c6fc1e49da8e Oct 10 15:07:09 crc kubenswrapper[4788]: I1010 15:07:09.689312 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerStarted","Data":"a28995f507205de0012a07bbf23fe5bdfc9d314e572c4f976638c6fc1e49da8e"} Oct 10 15:07:10 crc kubenswrapper[4788]: I1010 15:07:10.250604 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9801af7-97f4-42c6-bae6-9d617fda37ef" path="/var/lib/kubelet/pods/e9801af7-97f4-42c6-bae6-9d617fda37ef/volumes" Oct 10 15:07:10 crc kubenswrapper[4788]: I1010 15:07:10.707749 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerStarted","Data":"2d6153d79b9a0c23fdfaba336c12b1d4239ff99581357a3dddaa2a13e66e9499"} Oct 10 15:07:11 crc kubenswrapper[4788]: I1010 15:07:11.739990 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerStarted","Data":"4c3acbecf325173d28f79d929e2c02d58e460088b059fe2cd30d126c41591ab3"} Oct 10 15:07:11 crc kubenswrapper[4788]: I1010 15:07:11.744007 4788 generic.go:334] "Generic (PLEG): container finished" podID="bd483623-d8ec-4daa-9641-a25e3c92b0eb" containerID="1e94fd37cb99c18e9e70ea41a5932cd392fae9e72bcd73f02398b3f4d1d01fca" exitCode=0 Oct 10 15:07:11 crc kubenswrapper[4788]: I1010 15:07:11.744055 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-d65h8" event={"ID":"bd483623-d8ec-4daa-9641-a25e3c92b0eb","Type":"ContainerDied","Data":"1e94fd37cb99c18e9e70ea41a5932cd392fae9e72bcd73f02398b3f4d1d01fca"} Oct 10 15:07:12 crc kubenswrapper[4788]: I1010 15:07:12.758971 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerStarted","Data":"3d5409a1cf47943139f8aadab679681c2155735e3b31ea44ebf9c3f8a684c34a"} Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.138475 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.225888 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-combined-ca-bundle\") pod \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.226513 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7vfn\" (UniqueName: \"kubernetes.io/projected/bd483623-d8ec-4daa-9641-a25e3c92b0eb-kube-api-access-k7vfn\") pod \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.226563 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-config-data\") pod \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.226659 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-scripts\") pod \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\" (UID: \"bd483623-d8ec-4daa-9641-a25e3c92b0eb\") " Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.235052 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd483623-d8ec-4daa-9641-a25e3c92b0eb-kube-api-access-k7vfn" (OuterVolumeSpecName: "kube-api-access-k7vfn") pod "bd483623-d8ec-4daa-9641-a25e3c92b0eb" (UID: "bd483623-d8ec-4daa-9641-a25e3c92b0eb"). InnerVolumeSpecName "kube-api-access-k7vfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.246277 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-scripts" (OuterVolumeSpecName: "scripts") pod "bd483623-d8ec-4daa-9641-a25e3c92b0eb" (UID: "bd483623-d8ec-4daa-9641-a25e3c92b0eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.263329 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-config-data" (OuterVolumeSpecName: "config-data") pod "bd483623-d8ec-4daa-9641-a25e3c92b0eb" (UID: "bd483623-d8ec-4daa-9641-a25e3c92b0eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.263591 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd483623-d8ec-4daa-9641-a25e3c92b0eb" (UID: "bd483623-d8ec-4daa-9641-a25e3c92b0eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.328962 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.329238 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7vfn\" (UniqueName: \"kubernetes.io/projected/bd483623-d8ec-4daa-9641-a25e3c92b0eb-kube-api-access-k7vfn\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.329301 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.329356 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd483623-d8ec-4daa-9641-a25e3c92b0eb-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.767902 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-d65h8" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.767891 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-d65h8" event={"ID":"bd483623-d8ec-4daa-9641-a25e3c92b0eb","Type":"ContainerDied","Data":"011b50a5a5fce056fcc11d839aca55296d579d0126e3e0292618659150bbff97"} Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.768075 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="011b50a5a5fce056fcc11d839aca55296d579d0126e3e0292618659150bbff97" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.771723 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerStarted","Data":"11b44c1f6365a782b0b7cb7559f79e19ab5b11e060b32f8b117349ae62639bc0"} Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.772853 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.821362 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.849343943 podStartE2EDuration="5.821342267s" podCreationTimestamp="2025-10-10 15:07:08 +0000 UTC" firstStartedPulling="2025-10-10 15:07:09.418207088 +0000 UTC m=+1331.867922636" lastFinishedPulling="2025-10-10 15:07:13.390205402 +0000 UTC m=+1335.839920960" observedRunningTime="2025-10-10 15:07:13.797991907 +0000 UTC m=+1336.247707465" watchObservedRunningTime="2025-10-10 15:07:13.821342267 +0000 UTC m=+1336.271057805" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.901079 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 15:07:13 crc kubenswrapper[4788]: E1010 15:07:13.902150 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd483623-d8ec-4daa-9641-a25e3c92b0eb" containerName="nova-cell0-conductor-db-sync" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.902172 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd483623-d8ec-4daa-9641-a25e3c92b0eb" containerName="nova-cell0-conductor-db-sync" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.902357 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd483623-d8ec-4daa-9641-a25e3c92b0eb" containerName="nova-cell0-conductor-db-sync" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.902981 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.911441 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.911590 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-r6d9t" Oct 10 15:07:13 crc kubenswrapper[4788]: I1010 15:07:13.951825 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.048406 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.048802 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crw2j\" (UniqueName: \"kubernetes.io/projected/cf293818-85fb-46dc-8cee-ba8eca827bb7-kube-api-access-crw2j\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.048857 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.151019 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.151309 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.151475 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crw2j\" (UniqueName: \"kubernetes.io/projected/cf293818-85fb-46dc-8cee-ba8eca827bb7-kube-api-access-crw2j\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.157053 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.162670 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.173896 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crw2j\" (UniqueName: \"kubernetes.io/projected/cf293818-85fb-46dc-8cee-ba8eca827bb7-kube-api-access-crw2j\") pod \"nova-cell0-conductor-0\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.265332 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:14 crc kubenswrapper[4788]: W1010 15:07:14.767359 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf293818_85fb_46dc_8cee_ba8eca827bb7.slice/crio-6be74aa1cff12947235f803c8b52f306fe46a6930bef9bbe5e4b0dbce0cd3e05 WatchSource:0}: Error finding container 6be74aa1cff12947235f803c8b52f306fe46a6930bef9bbe5e4b0dbce0cd3e05: Status 404 returned error can't find the container with id 6be74aa1cff12947235f803c8b52f306fe46a6930bef9bbe5e4b0dbce0cd3e05 Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.776689 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 15:07:14 crc kubenswrapper[4788]: I1010 15:07:14.784085 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cf293818-85fb-46dc-8cee-ba8eca827bb7","Type":"ContainerStarted","Data":"6be74aa1cff12947235f803c8b52f306fe46a6930bef9bbe5e4b0dbce0cd3e05"} Oct 10 15:07:15 crc kubenswrapper[4788]: I1010 15:07:15.799134 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cf293818-85fb-46dc-8cee-ba8eca827bb7","Type":"ContainerStarted","Data":"1ac5b4a4856f6b0a772902a9c4e92e64c3a65186c5c7b198248ff40fef09742d"} Oct 10 15:07:15 crc kubenswrapper[4788]: I1010 15:07:15.799723 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:15 crc kubenswrapper[4788]: I1010 15:07:15.830295 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.830273529 podStartE2EDuration="2.830273529s" podCreationTimestamp="2025-10-10 15:07:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:15.81841223 +0000 UTC m=+1338.268127818" watchObservedRunningTime="2025-10-10 15:07:15.830273529 +0000 UTC m=+1338.279989077" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.303066 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.786688 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-gxptq"] Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.788011 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.789955 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.796290 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.806957 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gxptq"] Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.979240 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-scripts\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.979301 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk7jq\" (UniqueName: \"kubernetes.io/projected/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-kube-api-access-tk7jq\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.979335 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-config-data\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.979408 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.987971 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.989629 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.993705 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.997555 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:19 crc kubenswrapper[4788]: I1010 15:07:19.999719 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.003140 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.035282 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.057425 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.081559 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-scripts\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.081626 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-config-data\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.081649 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk7jq\" (UniqueName: \"kubernetes.io/projected/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-kube-api-access-tk7jq\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.081678 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.096909 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-scripts\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.098488 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.100768 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-config-data\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.113826 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.115757 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.119042 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.124258 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.135898 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk7jq\" (UniqueName: \"kubernetes.io/projected/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-kube-api-access-tk7jq\") pod \"nova-cell0-cell-mapping-gxptq\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.186603 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-config-data\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.186680 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c18dca-feb6-4006-a0ec-817e01a156c9-logs\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.186724 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.186753 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c9wt\" (UniqueName: \"kubernetes.io/projected/78c18dca-feb6-4006-a0ec-817e01a156c9-kube-api-access-7c9wt\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.186792 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.186832 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.186856 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7xbt\" (UniqueName: \"kubernetes.io/projected/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-kube-api-access-k7xbt\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.212329 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hhd8c"] Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.216156 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.223023 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.236870 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.244460 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.284703 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291303 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdv29\" (UniqueName: \"kubernetes.io/projected/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-kube-api-access-fdv29\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291394 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-config-data\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291425 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c18dca-feb6-4006-a0ec-817e01a156c9-logs\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291461 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291491 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c9wt\" (UniqueName: \"kubernetes.io/projected/78c18dca-feb6-4006-a0ec-817e01a156c9-kube-api-access-7c9wt\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291519 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291545 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-logs\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291563 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-config-data\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291585 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291630 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.291664 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7xbt\" (UniqueName: \"kubernetes.io/projected/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-kube-api-access-k7xbt\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.293040 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c18dca-feb6-4006-a0ec-817e01a156c9-logs\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.293728 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hhd8c"] Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.298102 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.303107 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.303758 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.304310 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-config-data\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.314318 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c9wt\" (UniqueName: \"kubernetes.io/projected/78c18dca-feb6-4006-a0ec-817e01a156c9-kube-api-access-7c9wt\") pod \"nova-api-0\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.328642 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7xbt\" (UniqueName: \"kubernetes.io/projected/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-kube-api-access-k7xbt\") pod \"nova-cell1-novncproxy-0\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.335662 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.351421 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.397705 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdv29\" (UniqueName: \"kubernetes.io/projected/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-kube-api-access-fdv29\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.397763 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frw2b\" (UniqueName: \"kubernetes.io/projected/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-kube-api-access-frw2b\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.397844 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.397889 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.397936 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.397982 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-config\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.398024 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.398095 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.398138 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-logs\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.398169 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-config-data\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.398184 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-config-data\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.398212 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnxtj\" (UniqueName: \"kubernetes.io/projected/f3f7a671-e17f-4be0-8afb-d10395da7ad8-kube-api-access-fnxtj\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.398274 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.401980 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-logs\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.418074 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.418670 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-config-data\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.419491 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.437763 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdv29\" (UniqueName: \"kubernetes.io/projected/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-kube-api-access-fdv29\") pod \"nova-metadata-0\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.501764 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-config-data\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.502186 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnxtj\" (UniqueName: \"kubernetes.io/projected/f3f7a671-e17f-4be0-8afb-d10395da7ad8-kube-api-access-fnxtj\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.502237 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.502299 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frw2b\" (UniqueName: \"kubernetes.io/projected/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-kube-api-access-frw2b\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.502349 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.502378 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.502410 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.502449 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-config\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.502487 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.504443 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.509087 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.510236 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.510423 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-config\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.511982 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-config-data\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.513935 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.521396 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.525931 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.535999 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frw2b\" (UniqueName: \"kubernetes.io/projected/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-kube-api-access-frw2b\") pod \"dnsmasq-dns-845d6d6f59-hhd8c\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.557244 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.577620 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnxtj\" (UniqueName: \"kubernetes.io/projected/f3f7a671-e17f-4be0-8afb-d10395da7ad8-kube-api-access-fnxtj\") pod \"nova-scheduler-0\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:20 crc kubenswrapper[4788]: I1010 15:07:20.637542 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.039958 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.193035 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g5tcl"] Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.195223 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.197125 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.204946 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.218780 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g5tcl"] Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.261144 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.288797 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-scripts\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.288859 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.288888 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsxwv\" (UniqueName: \"kubernetes.io/projected/64635416-9e49-4b2b-8b68-4ed756fbb05c-kube-api-access-zsxwv\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.288990 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-config-data\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.391291 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hhd8c"] Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.392015 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-config-data\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.392141 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-scripts\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.392199 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.392227 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsxwv\" (UniqueName: \"kubernetes.io/projected/64635416-9e49-4b2b-8b68-4ed756fbb05c-kube-api-access-zsxwv\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.406089 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-scripts\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.412579 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-config-data\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.412720 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.425307 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsxwv\" (UniqueName: \"kubernetes.io/projected/64635416-9e49-4b2b-8b68-4ed756fbb05c-kube-api-access-zsxwv\") pod \"nova-cell1-conductor-db-sync-g5tcl\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.425387 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:21 crc kubenswrapper[4788]: W1010 15:07:21.432261 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddccfc0c7_923d_4fc6_b037_91d85d6dcae1.slice/crio-f4715ae7e426e57f046f5f8fe9a63620165e14c8f508895b2a2c61a5d4d13ca1 WatchSource:0}: Error finding container f4715ae7e426e57f046f5f8fe9a63620165e14c8f508895b2a2c61a5d4d13ca1: Status 404 returned error can't find the container with id f4715ae7e426e57f046f5f8fe9a63620165e14c8f508895b2a2c61a5d4d13ca1 Oct 10 15:07:21 crc kubenswrapper[4788]: W1010 15:07:21.433588 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b98f8e7_0f9c_44f7_9b8b_0a5638c0875d.slice/crio-e199d6a2467b69f8c5f728be607c2efdee8cacac218847628272c973313eec5e WatchSource:0}: Error finding container e199d6a2467b69f8c5f728be607c2efdee8cacac218847628272c973313eec5e: Status 404 returned error can't find the container with id e199d6a2467b69f8c5f728be607c2efdee8cacac218847628272c973313eec5e Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.435510 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gxptq"] Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.446644 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.514969 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.918865 4788 generic.go:334] "Generic (PLEG): container finished" podID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" containerID="a27b71c2742a9f61882af118971ebd23910a6cecfff6ae21668a4ef35df711ed" exitCode=0 Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.919055 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" event={"ID":"c40f46e4-4197-4b2f-9cf7-c3b3d013838d","Type":"ContainerDied","Data":"a27b71c2742a9f61882af118971ebd23910a6cecfff6ae21668a4ef35df711ed"} Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.919517 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" event={"ID":"c40f46e4-4197-4b2f-9cf7-c3b3d013838d","Type":"ContainerStarted","Data":"9867e85b57b7659573a6b044235801cdfedbecd6e5a2d59197e8d5b562255a6c"} Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.922065 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f3f7a671-e17f-4be0-8afb-d10395da7ad8","Type":"ContainerStarted","Data":"67f1bdd906e2aa86153e2ca8c12bce6176cc913f19d7d1bbf378a6ec716f4dd0"} Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.927794 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dccfc0c7-923d-4fc6-b037-91d85d6dcae1","Type":"ContainerStarted","Data":"f4715ae7e426e57f046f5f8fe9a63620165e14c8f508895b2a2c61a5d4d13ca1"} Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.930807 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f0bdf94-e96f-47ca-96a6-677a9546fc4d","Type":"ContainerStarted","Data":"3d17c2ebcd07aad38bba0af6889885d35ea62ce5db7407a74accad7dec18861c"} Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.932465 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gxptq" event={"ID":"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d","Type":"ContainerStarted","Data":"51707ce7890bf73c148c3998b37c2bd349e7afe315f7fbe5508284b03085928d"} Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.932506 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gxptq" event={"ID":"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d","Type":"ContainerStarted","Data":"e199d6a2467b69f8c5f728be607c2efdee8cacac218847628272c973313eec5e"} Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.934788 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c18dca-feb6-4006-a0ec-817e01a156c9","Type":"ContainerStarted","Data":"d6e0994090b1b521e2bfa84d683e5b56386f29fe2b62f7256841b359b7fd4a95"} Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.981644 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-gxptq" podStartSLOduration=2.981621739 podStartE2EDuration="2.981621739s" podCreationTimestamp="2025-10-10 15:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:21.966737147 +0000 UTC m=+1344.416452695" watchObservedRunningTime="2025-10-10 15:07:21.981621739 +0000 UTC m=+1344.431337287" Oct 10 15:07:21 crc kubenswrapper[4788]: I1010 15:07:21.997140 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g5tcl"] Oct 10 15:07:22 crc kubenswrapper[4788]: W1010 15:07:22.004009 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64635416_9e49_4b2b_8b68_4ed756fbb05c.slice/crio-4d8c917c019c28c12be1f2d17f0be1ffc177cf95d695decc1cd5bc7b2fa90dfd WatchSource:0}: Error finding container 4d8c917c019c28c12be1f2d17f0be1ffc177cf95d695decc1cd5bc7b2fa90dfd: Status 404 returned error can't find the container with id 4d8c917c019c28c12be1f2d17f0be1ffc177cf95d695decc1cd5bc7b2fa90dfd Oct 10 15:07:22 crc kubenswrapper[4788]: I1010 15:07:22.961713 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" event={"ID":"64635416-9e49-4b2b-8b68-4ed756fbb05c","Type":"ContainerStarted","Data":"9c4939632ebd89196bb2dfa61693556ec6deaee5d6432afbfdb9352a9b188e2a"} Oct 10 15:07:22 crc kubenswrapper[4788]: I1010 15:07:22.962204 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" event={"ID":"64635416-9e49-4b2b-8b68-4ed756fbb05c","Type":"ContainerStarted","Data":"4d8c917c019c28c12be1f2d17f0be1ffc177cf95d695decc1cd5bc7b2fa90dfd"} Oct 10 15:07:22 crc kubenswrapper[4788]: I1010 15:07:22.977413 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" event={"ID":"c40f46e4-4197-4b2f-9cf7-c3b3d013838d","Type":"ContainerStarted","Data":"6d0cb49f9707675944733cc27f4c654b8f06a53b69f92ad3276a8956de8ab293"} Oct 10 15:07:22 crc kubenswrapper[4788]: I1010 15:07:22.977461 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:22 crc kubenswrapper[4788]: I1010 15:07:22.987923 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" podStartSLOduration=1.987905888 podStartE2EDuration="1.987905888s" podCreationTimestamp="2025-10-10 15:07:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:22.98648625 +0000 UTC m=+1345.436201798" watchObservedRunningTime="2025-10-10 15:07:22.987905888 +0000 UTC m=+1345.437621436" Oct 10 15:07:23 crc kubenswrapper[4788]: I1010 15:07:23.008894 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" podStartSLOduration=3.008867894 podStartE2EDuration="3.008867894s" podCreationTimestamp="2025-10-10 15:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:23.006947183 +0000 UTC m=+1345.456662731" watchObservedRunningTime="2025-10-10 15:07:23.008867894 +0000 UTC m=+1345.458583442" Oct 10 15:07:24 crc kubenswrapper[4788]: I1010 15:07:24.443879 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:24 crc kubenswrapper[4788]: I1010 15:07:24.451184 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:25 crc kubenswrapper[4788]: I1010 15:07:25.003154 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f3f7a671-e17f-4be0-8afb-d10395da7ad8","Type":"ContainerStarted","Data":"155460bd9f7f7a8b06685eb2550e5f97ae4172a9cd51c3ea8ed6a80fa5b5287e"} Oct 10 15:07:25 crc kubenswrapper[4788]: I1010 15:07:25.011906 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f0bdf94-e96f-47ca-96a6-677a9546fc4d","Type":"ContainerStarted","Data":"226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb"} Oct 10 15:07:25 crc kubenswrapper[4788]: I1010 15:07:25.012153 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3f0bdf94-e96f-47ca-96a6-677a9546fc4d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb" gracePeriod=30 Oct 10 15:07:25 crc kubenswrapper[4788]: I1010 15:07:25.021091 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c18dca-feb6-4006-a0ec-817e01a156c9","Type":"ContainerStarted","Data":"3e0368d97024ca12075d24fda31d7d4eead57732acfefdcbd6d20c33d4f7f6f7"} Oct 10 15:07:25 crc kubenswrapper[4788]: I1010 15:07:25.031360 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.871861274 podStartE2EDuration="5.031344423s" podCreationTimestamp="2025-10-10 15:07:20 +0000 UTC" firstStartedPulling="2025-10-10 15:07:21.41820249 +0000 UTC m=+1343.867918038" lastFinishedPulling="2025-10-10 15:07:24.577685639 +0000 UTC m=+1347.027401187" observedRunningTime="2025-10-10 15:07:25.021096955 +0000 UTC m=+1347.470812503" watchObservedRunningTime="2025-10-10 15:07:25.031344423 +0000 UTC m=+1347.481059971" Oct 10 15:07:25 crc kubenswrapper[4788]: I1010 15:07:25.044344 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5110003069999998 podStartE2EDuration="6.044310443s" podCreationTimestamp="2025-10-10 15:07:19 +0000 UTC" firstStartedPulling="2025-10-10 15:07:21.046018497 +0000 UTC m=+1343.495734045" lastFinishedPulling="2025-10-10 15:07:24.579328633 +0000 UTC m=+1347.029044181" observedRunningTime="2025-10-10 15:07:25.039437081 +0000 UTC m=+1347.489152629" watchObservedRunningTime="2025-10-10 15:07:25.044310443 +0000 UTC m=+1347.494025991" Oct 10 15:07:25 crc kubenswrapper[4788]: I1010 15:07:25.336332 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:25 crc kubenswrapper[4788]: I1010 15:07:25.639284 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.036185 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dccfc0c7-923d-4fc6-b037-91d85d6dcae1","Type":"ContainerStarted","Data":"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de"} Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.036256 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dccfc0c7-923d-4fc6-b037-91d85d6dcae1","Type":"ContainerStarted","Data":"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5"} Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.036250 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerName="nova-metadata-log" containerID="cri-o://420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5" gracePeriod=30 Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.036324 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerName="nova-metadata-metadata" containerID="cri-o://2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de" gracePeriod=30 Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.045321 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c18dca-feb6-4006-a0ec-817e01a156c9","Type":"ContainerStarted","Data":"4104f4cfee1edb339a3c5824fe82691a528e1cc74ee8a2ec60a5c133c43e91ee"} Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.062506 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.919141502 podStartE2EDuration="6.062484294s" podCreationTimestamp="2025-10-10 15:07:20 +0000 UTC" firstStartedPulling="2025-10-10 15:07:21.435415635 +0000 UTC m=+1343.885131183" lastFinishedPulling="2025-10-10 15:07:24.578758407 +0000 UTC m=+1347.028473975" observedRunningTime="2025-10-10 15:07:26.057730336 +0000 UTC m=+1348.507445884" watchObservedRunningTime="2025-10-10 15:07:26.062484294 +0000 UTC m=+1348.512199852" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.082624 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.765522113 podStartE2EDuration="7.082603448s" podCreationTimestamp="2025-10-10 15:07:19 +0000 UTC" firstStartedPulling="2025-10-10 15:07:21.266236416 +0000 UTC m=+1343.715951964" lastFinishedPulling="2025-10-10 15:07:24.583317751 +0000 UTC m=+1347.033033299" observedRunningTime="2025-10-10 15:07:26.078441265 +0000 UTC m=+1348.528156833" watchObservedRunningTime="2025-10-10 15:07:26.082603448 +0000 UTC m=+1348.532318996" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.704519 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.720874 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-combined-ca-bundle\") pod \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.720948 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-config-data\") pod \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.720986 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdv29\" (UniqueName: \"kubernetes.io/projected/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-kube-api-access-fdv29\") pod \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.721006 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-logs\") pod \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\" (UID: \"dccfc0c7-923d-4fc6-b037-91d85d6dcae1\") " Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.721602 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-logs" (OuterVolumeSpecName: "logs") pod "dccfc0c7-923d-4fc6-b037-91d85d6dcae1" (UID: "dccfc0c7-923d-4fc6-b037-91d85d6dcae1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.737657 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-kube-api-access-fdv29" (OuterVolumeSpecName: "kube-api-access-fdv29") pod "dccfc0c7-923d-4fc6-b037-91d85d6dcae1" (UID: "dccfc0c7-923d-4fc6-b037-91d85d6dcae1"). InnerVolumeSpecName "kube-api-access-fdv29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.782512 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-config-data" (OuterVolumeSpecName: "config-data") pod "dccfc0c7-923d-4fc6-b037-91d85d6dcae1" (UID: "dccfc0c7-923d-4fc6-b037-91d85d6dcae1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.832127 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.832216 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdv29\" (UniqueName: \"kubernetes.io/projected/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-kube-api-access-fdv29\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.832252 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.833747 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dccfc0c7-923d-4fc6-b037-91d85d6dcae1" (UID: "dccfc0c7-923d-4fc6-b037-91d85d6dcae1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:26 crc kubenswrapper[4788]: I1010 15:07:26.934938 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dccfc0c7-923d-4fc6-b037-91d85d6dcae1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.055010 4788 generic.go:334] "Generic (PLEG): container finished" podID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerID="2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de" exitCode=0 Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.055058 4788 generic.go:334] "Generic (PLEG): container finished" podID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerID="420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5" exitCode=143 Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.056448 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.056565 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dccfc0c7-923d-4fc6-b037-91d85d6dcae1","Type":"ContainerDied","Data":"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de"} Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.056606 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dccfc0c7-923d-4fc6-b037-91d85d6dcae1","Type":"ContainerDied","Data":"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5"} Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.056624 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"dccfc0c7-923d-4fc6-b037-91d85d6dcae1","Type":"ContainerDied","Data":"f4715ae7e426e57f046f5f8fe9a63620165e14c8f508895b2a2c61a5d4d13ca1"} Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.056643 4788 scope.go:117] "RemoveContainer" containerID="2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.084298 4788 scope.go:117] "RemoveContainer" containerID="420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.093267 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.110198 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.121364 4788 scope.go:117] "RemoveContainer" containerID="2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de" Oct 10 15:07:27 crc kubenswrapper[4788]: E1010 15:07:27.121854 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de\": container with ID starting with 2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de not found: ID does not exist" containerID="2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.121910 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de"} err="failed to get container status \"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de\": rpc error: code = NotFound desc = could not find container \"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de\": container with ID starting with 2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de not found: ID does not exist" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.121939 4788 scope.go:117] "RemoveContainer" containerID="420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5" Oct 10 15:07:27 crc kubenswrapper[4788]: E1010 15:07:27.122223 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5\": container with ID starting with 420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5 not found: ID does not exist" containerID="420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.122243 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5"} err="failed to get container status \"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5\": rpc error: code = NotFound desc = could not find container \"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5\": container with ID starting with 420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5 not found: ID does not exist" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.122256 4788 scope.go:117] "RemoveContainer" containerID="2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.122492 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de"} err="failed to get container status \"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de\": rpc error: code = NotFound desc = could not find container \"2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de\": container with ID starting with 2f294afe06582554aca6d5ef846c792280b966993dc98535a86053cde12c45de not found: ID does not exist" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.122509 4788 scope.go:117] "RemoveContainer" containerID="420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.122718 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5"} err="failed to get container status \"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5\": rpc error: code = NotFound desc = could not find container \"420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5\": container with ID starting with 420a7eec91967d38864e071b52d45b253b55ee8491eaf924a326d1ed130dfcb5 not found: ID does not exist" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.125167 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:27 crc kubenswrapper[4788]: E1010 15:07:27.125801 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerName="nova-metadata-log" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.125822 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerName="nova-metadata-log" Oct 10 15:07:27 crc kubenswrapper[4788]: E1010 15:07:27.125854 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerName="nova-metadata-metadata" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.125861 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerName="nova-metadata-metadata" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.126086 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerName="nova-metadata-metadata" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.126109 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" containerName="nova-metadata-log" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.127368 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.134489 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.134796 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.139116 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dphvq\" (UniqueName: \"kubernetes.io/projected/c62ef92f-3981-4eca-b119-faeb979af999-kube-api-access-dphvq\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.139183 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.139211 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.139302 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c62ef92f-3981-4eca-b119-faeb979af999-logs\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.139382 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-config-data\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.158969 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.242295 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dphvq\" (UniqueName: \"kubernetes.io/projected/c62ef92f-3981-4eca-b119-faeb979af999-kube-api-access-dphvq\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.242380 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.242412 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.242681 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c62ef92f-3981-4eca-b119-faeb979af999-logs\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.243572 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-config-data\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.243573 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c62ef92f-3981-4eca-b119-faeb979af999-logs\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.248759 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.252393 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.252800 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-config-data\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.260939 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dphvq\" (UniqueName: \"kubernetes.io/projected/c62ef92f-3981-4eca-b119-faeb979af999-kube-api-access-dphvq\") pod \"nova-metadata-0\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " pod="openstack/nova-metadata-0" Oct 10 15:07:27 crc kubenswrapper[4788]: I1010 15:07:27.462952 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:28 crc kubenswrapper[4788]: I1010 15:07:28.010167 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:28 crc kubenswrapper[4788]: I1010 15:07:28.073703 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c62ef92f-3981-4eca-b119-faeb979af999","Type":"ContainerStarted","Data":"10735c9b0b91c22c3064d2ecebbc95d31ef896191979f78778192cd84a2cd6ea"} Oct 10 15:07:28 crc kubenswrapper[4788]: I1010 15:07:28.249236 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dccfc0c7-923d-4fc6-b037-91d85d6dcae1" path="/var/lib/kubelet/pods/dccfc0c7-923d-4fc6-b037-91d85d6dcae1/volumes" Oct 10 15:07:29 crc kubenswrapper[4788]: I1010 15:07:29.092622 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c62ef92f-3981-4eca-b119-faeb979af999","Type":"ContainerStarted","Data":"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d"} Oct 10 15:07:29 crc kubenswrapper[4788]: I1010 15:07:29.093333 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c62ef92f-3981-4eca-b119-faeb979af999","Type":"ContainerStarted","Data":"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40"} Oct 10 15:07:29 crc kubenswrapper[4788]: I1010 15:07:29.116005 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.11597384 podStartE2EDuration="2.11597384s" podCreationTimestamp="2025-10-10 15:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:29.114034048 +0000 UTC m=+1351.563749596" watchObservedRunningTime="2025-10-10 15:07:29.11597384 +0000 UTC m=+1351.565689388" Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.127270 4788 generic.go:334] "Generic (PLEG): container finished" podID="64635416-9e49-4b2b-8b68-4ed756fbb05c" containerID="9c4939632ebd89196bb2dfa61693556ec6deaee5d6432afbfdb9352a9b188e2a" exitCode=0 Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.127441 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" event={"ID":"64635416-9e49-4b2b-8b68-4ed756fbb05c","Type":"ContainerDied","Data":"9c4939632ebd89196bb2dfa61693556ec6deaee5d6432afbfdb9352a9b188e2a"} Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.146195 4788 generic.go:334] "Generic (PLEG): container finished" podID="5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" containerID="51707ce7890bf73c148c3998b37c2bd349e7afe315f7fbe5508284b03085928d" exitCode=0 Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.146662 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gxptq" event={"ID":"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d","Type":"ContainerDied","Data":"51707ce7890bf73c148c3998b37c2bd349e7afe315f7fbe5508284b03085928d"} Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.352456 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.353189 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.564318 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.633734 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5bj6f"] Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.634263 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" podUID="8f684fc3-b2c2-454a-b70e-5a19f9520710" containerName="dnsmasq-dns" containerID="cri-o://a2a9700b17de8390730939a608c45fe4b3b60e9b1e67395aceadf7ac2576f14f" gracePeriod=10 Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.639584 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 15:07:30 crc kubenswrapper[4788]: I1010 15:07:30.688373 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.166202 4788 generic.go:334] "Generic (PLEG): container finished" podID="8f684fc3-b2c2-454a-b70e-5a19f9520710" containerID="a2a9700b17de8390730939a608c45fe4b3b60e9b1e67395aceadf7ac2576f14f" exitCode=0 Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.166575 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" event={"ID":"8f684fc3-b2c2-454a-b70e-5a19f9520710","Type":"ContainerDied","Data":"a2a9700b17de8390730939a608c45fe4b3b60e9b1e67395aceadf7ac2576f14f"} Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.168286 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" event={"ID":"8f684fc3-b2c2-454a-b70e-5a19f9520710","Type":"ContainerDied","Data":"51aea72c498a2a95988c0dc5418c3d7f01427762774eaf668f2fb9c27c3851b3"} Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.168319 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51aea72c498a2a95988c0dc5418c3d7f01427762774eaf668f2fb9c27c3851b3" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.277368 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.294339 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.443280 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.443593 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.476244 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-swift-storage-0\") pod \"8f684fc3-b2c2-454a-b70e-5a19f9520710\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.476580 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws82x\" (UniqueName: \"kubernetes.io/projected/8f684fc3-b2c2-454a-b70e-5a19f9520710-kube-api-access-ws82x\") pod \"8f684fc3-b2c2-454a-b70e-5a19f9520710\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.476710 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-svc\") pod \"8f684fc3-b2c2-454a-b70e-5a19f9520710\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.476867 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-nb\") pod \"8f684fc3-b2c2-454a-b70e-5a19f9520710\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.478244 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-config\") pod \"8f684fc3-b2c2-454a-b70e-5a19f9520710\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.478380 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-sb\") pod \"8f684fc3-b2c2-454a-b70e-5a19f9520710\" (UID: \"8f684fc3-b2c2-454a-b70e-5a19f9520710\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.540794 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f684fc3-b2c2-454a-b70e-5a19f9520710-kube-api-access-ws82x" (OuterVolumeSpecName: "kube-api-access-ws82x") pod "8f684fc3-b2c2-454a-b70e-5a19f9520710" (UID: "8f684fc3-b2c2-454a-b70e-5a19f9520710"). InnerVolumeSpecName "kube-api-access-ws82x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.582850 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws82x\" (UniqueName: \"kubernetes.io/projected/8f684fc3-b2c2-454a-b70e-5a19f9520710-kube-api-access-ws82x\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.630832 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-config" (OuterVolumeSpecName: "config") pod "8f684fc3-b2c2-454a-b70e-5a19f9520710" (UID: "8f684fc3-b2c2-454a-b70e-5a19f9520710"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.649116 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8f684fc3-b2c2-454a-b70e-5a19f9520710" (UID: "8f684fc3-b2c2-454a-b70e-5a19f9520710"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.684627 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.684664 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.703743 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.711926 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8f684fc3-b2c2-454a-b70e-5a19f9520710" (UID: "8f684fc3-b2c2-454a-b70e-5a19f9520710"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.713730 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8f684fc3-b2c2-454a-b70e-5a19f9520710" (UID: "8f684fc3-b2c2-454a-b70e-5a19f9520710"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.720092 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8f684fc3-b2c2-454a-b70e-5a19f9520710" (UID: "8f684fc3-b2c2-454a-b70e-5a19f9520710"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.788105 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-config-data\") pod \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.788692 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-scripts\") pod \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.788797 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk7jq\" (UniqueName: \"kubernetes.io/projected/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-kube-api-access-tk7jq\") pod \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.788871 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-combined-ca-bundle\") pod \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\" (UID: \"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d\") " Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.789615 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.789640 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.789665 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f684fc3-b2c2-454a-b70e-5a19f9520710-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.800360 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-scripts" (OuterVolumeSpecName: "scripts") pod "5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" (UID: "5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.801109 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-kube-api-access-tk7jq" (OuterVolumeSpecName: "kube-api-access-tk7jq") pod "5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" (UID: "5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d"). InnerVolumeSpecName "kube-api-access-tk7jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.842175 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-config-data" (OuterVolumeSpecName: "config-data") pod "5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" (UID: "5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.851264 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" (UID: "5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.891879 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.891909 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.891935 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk7jq\" (UniqueName: \"kubernetes.io/projected/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-kube-api-access-tk7jq\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.891948 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:31 crc kubenswrapper[4788]: I1010 15:07:31.927205 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.094940 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-config-data\") pod \"64635416-9e49-4b2b-8b68-4ed756fbb05c\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.096470 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsxwv\" (UniqueName: \"kubernetes.io/projected/64635416-9e49-4b2b-8b68-4ed756fbb05c-kube-api-access-zsxwv\") pod \"64635416-9e49-4b2b-8b68-4ed756fbb05c\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.096506 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-scripts\") pod \"64635416-9e49-4b2b-8b68-4ed756fbb05c\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.097329 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-combined-ca-bundle\") pod \"64635416-9e49-4b2b-8b68-4ed756fbb05c\" (UID: \"64635416-9e49-4b2b-8b68-4ed756fbb05c\") " Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.100923 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64635416-9e49-4b2b-8b68-4ed756fbb05c-kube-api-access-zsxwv" (OuterVolumeSpecName: "kube-api-access-zsxwv") pod "64635416-9e49-4b2b-8b68-4ed756fbb05c" (UID: "64635416-9e49-4b2b-8b68-4ed756fbb05c"). InnerVolumeSpecName "kube-api-access-zsxwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.118659 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-scripts" (OuterVolumeSpecName: "scripts") pod "64635416-9e49-4b2b-8b68-4ed756fbb05c" (UID: "64635416-9e49-4b2b-8b68-4ed756fbb05c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.128938 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64635416-9e49-4b2b-8b68-4ed756fbb05c" (UID: "64635416-9e49-4b2b-8b68-4ed756fbb05c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.129552 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-config-data" (OuterVolumeSpecName: "config-data") pod "64635416-9e49-4b2b-8b68-4ed756fbb05c" (UID: "64635416-9e49-4b2b-8b68-4ed756fbb05c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.187670 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gxptq" event={"ID":"5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d","Type":"ContainerDied","Data":"e199d6a2467b69f8c5f728be607c2efdee8cacac218847628272c973313eec5e"} Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.187730 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e199d6a2467b69f8c5f728be607c2efdee8cacac218847628272c973313eec5e" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.187827 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gxptq" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.203219 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.204671 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-g5tcl" event={"ID":"64635416-9e49-4b2b-8b68-4ed756fbb05c","Type":"ContainerDied","Data":"4d8c917c019c28c12be1f2d17f0be1ffc177cf95d695decc1cd5bc7b2fa90dfd"} Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.204729 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d8c917c019c28c12be1f2d17f0be1ffc177cf95d695decc1cd5bc7b2fa90dfd" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.204899 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-5bj6f" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.210433 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.210760 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsxwv\" (UniqueName: \"kubernetes.io/projected/64635416-9e49-4b2b-8b68-4ed756fbb05c-kube-api-access-zsxwv\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.210960 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.210987 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64635416-9e49-4b2b-8b68-4ed756fbb05c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.316158 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 15:07:32 crc kubenswrapper[4788]: E1010 15:07:32.317320 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f684fc3-b2c2-454a-b70e-5a19f9520710" containerName="dnsmasq-dns" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.317355 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f684fc3-b2c2-454a-b70e-5a19f9520710" containerName="dnsmasq-dns" Oct 10 15:07:32 crc kubenswrapper[4788]: E1010 15:07:32.317407 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64635416-9e49-4b2b-8b68-4ed756fbb05c" containerName="nova-cell1-conductor-db-sync" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.317418 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="64635416-9e49-4b2b-8b68-4ed756fbb05c" containerName="nova-cell1-conductor-db-sync" Oct 10 15:07:32 crc kubenswrapper[4788]: E1010 15:07:32.317461 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" containerName="nova-manage" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.320082 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" containerName="nova-manage" Oct 10 15:07:32 crc kubenswrapper[4788]: E1010 15:07:32.320165 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f684fc3-b2c2-454a-b70e-5a19f9520710" containerName="init" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.320175 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f684fc3-b2c2-454a-b70e-5a19f9520710" containerName="init" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.321117 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f684fc3-b2c2-454a-b70e-5a19f9520710" containerName="dnsmasq-dns" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.321199 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="64635416-9e49-4b2b-8b68-4ed756fbb05c" containerName="nova-cell1-conductor-db-sync" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.323345 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" containerName="nova-manage" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.325059 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.330304 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.349265 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.366080 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5bj6f"] Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.374197 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-5bj6f"] Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.414658 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.414985 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-log" containerID="cri-o://3e0368d97024ca12075d24fda31d7d4eead57732acfefdcbd6d20c33d4f7f6f7" gracePeriod=30 Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.415157 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-api" containerID="cri-o://4104f4cfee1edb339a3c5824fe82691a528e1cc74ee8a2ec60a5c133c43e91ee" gracePeriod=30 Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.435345 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.435670 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c62ef92f-3981-4eca-b119-faeb979af999" containerName="nova-metadata-log" containerID="cri-o://ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40" gracePeriod=30 Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.436559 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c62ef92f-3981-4eca-b119-faeb979af999" containerName="nova-metadata-metadata" containerID="cri-o://de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d" gracePeriod=30 Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.463243 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.463333 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.519442 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.519705 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvh8t\" (UniqueName: \"kubernetes.io/projected/02360792-0295-42be-a94c-de0e7d2aa376-kube-api-access-wvh8t\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.519798 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.538439 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.621675 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvh8t\" (UniqueName: \"kubernetes.io/projected/02360792-0295-42be-a94c-de0e7d2aa376-kube-api-access-wvh8t\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.621748 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.621816 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.627345 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.627434 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.641134 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvh8t\" (UniqueName: \"kubernetes.io/projected/02360792-0295-42be-a94c-de0e7d2aa376-kube-api-access-wvh8t\") pod \"nova-cell1-conductor-0\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:32 crc kubenswrapper[4788]: I1010 15:07:32.649565 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.097640 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.211689 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.231508 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c62ef92f-3981-4eca-b119-faeb979af999-logs\") pod \"c62ef92f-3981-4eca-b119-faeb979af999\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.231567 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-combined-ca-bundle\") pod \"c62ef92f-3981-4eca-b119-faeb979af999\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.231596 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-nova-metadata-tls-certs\") pod \"c62ef92f-3981-4eca-b119-faeb979af999\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.231639 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-config-data\") pod \"c62ef92f-3981-4eca-b119-faeb979af999\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.231686 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dphvq\" (UniqueName: \"kubernetes.io/projected/c62ef92f-3981-4eca-b119-faeb979af999-kube-api-access-dphvq\") pod \"c62ef92f-3981-4eca-b119-faeb979af999\" (UID: \"c62ef92f-3981-4eca-b119-faeb979af999\") " Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.239747 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c62ef92f-3981-4eca-b119-faeb979af999-kube-api-access-dphvq" (OuterVolumeSpecName: "kube-api-access-dphvq") pod "c62ef92f-3981-4eca-b119-faeb979af999" (UID: "c62ef92f-3981-4eca-b119-faeb979af999"). InnerVolumeSpecName "kube-api-access-dphvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.241968 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c62ef92f-3981-4eca-b119-faeb979af999-logs" (OuterVolumeSpecName: "logs") pod "c62ef92f-3981-4eca-b119-faeb979af999" (UID: "c62ef92f-3981-4eca-b119-faeb979af999"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.242272 4788 generic.go:334] "Generic (PLEG): container finished" podID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerID="3e0368d97024ca12075d24fda31d7d4eead57732acfefdcbd6d20c33d4f7f6f7" exitCode=143 Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.242403 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c18dca-feb6-4006-a0ec-817e01a156c9","Type":"ContainerDied","Data":"3e0368d97024ca12075d24fda31d7d4eead57732acfefdcbd6d20c33d4f7f6f7"} Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.272720 4788 generic.go:334] "Generic (PLEG): container finished" podID="c62ef92f-3981-4eca-b119-faeb979af999" containerID="de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d" exitCode=0 Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.272774 4788 generic.go:334] "Generic (PLEG): container finished" podID="c62ef92f-3981-4eca-b119-faeb979af999" containerID="ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40" exitCode=143 Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.272852 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.272961 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c62ef92f-3981-4eca-b119-faeb979af999","Type":"ContainerDied","Data":"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d"} Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.273014 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c62ef92f-3981-4eca-b119-faeb979af999","Type":"ContainerDied","Data":"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40"} Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.273025 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c62ef92f-3981-4eca-b119-faeb979af999","Type":"ContainerDied","Data":"10735c9b0b91c22c3064d2ecebbc95d31ef896191979f78778192cd84a2cd6ea"} Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.273044 4788 scope.go:117] "RemoveContainer" containerID="de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.274162 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f3f7a671-e17f-4be0-8afb-d10395da7ad8" containerName="nova-scheduler-scheduler" containerID="cri-o://155460bd9f7f7a8b06685eb2550e5f97ae4172a9cd51c3ea8ed6a80fa5b5287e" gracePeriod=30 Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.288265 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-config-data" (OuterVolumeSpecName: "config-data") pod "c62ef92f-3981-4eca-b119-faeb979af999" (UID: "c62ef92f-3981-4eca-b119-faeb979af999"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.295283 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c62ef92f-3981-4eca-b119-faeb979af999" (UID: "c62ef92f-3981-4eca-b119-faeb979af999"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.297625 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c62ef92f-3981-4eca-b119-faeb979af999" (UID: "c62ef92f-3981-4eca-b119-faeb979af999"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.333801 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c62ef92f-3981-4eca-b119-faeb979af999-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.333840 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.333851 4788 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.333860 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c62ef92f-3981-4eca-b119-faeb979af999-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.333869 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dphvq\" (UniqueName: \"kubernetes.io/projected/c62ef92f-3981-4eca-b119-faeb979af999-kube-api-access-dphvq\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.334054 4788 scope.go:117] "RemoveContainer" containerID="ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.362649 4788 scope.go:117] "RemoveContainer" containerID="de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d" Oct 10 15:07:33 crc kubenswrapper[4788]: E1010 15:07:33.365055 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d\": container with ID starting with de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d not found: ID does not exist" containerID="de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.365105 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d"} err="failed to get container status \"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d\": rpc error: code = NotFound desc = could not find container \"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d\": container with ID starting with de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d not found: ID does not exist" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.365134 4788 scope.go:117] "RemoveContainer" containerID="ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40" Oct 10 15:07:33 crc kubenswrapper[4788]: E1010 15:07:33.365550 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40\": container with ID starting with ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40 not found: ID does not exist" containerID="ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.365573 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40"} err="failed to get container status \"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40\": rpc error: code = NotFound desc = could not find container \"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40\": container with ID starting with ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40 not found: ID does not exist" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.365586 4788 scope.go:117] "RemoveContainer" containerID="de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.366969 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d"} err="failed to get container status \"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d\": rpc error: code = NotFound desc = could not find container \"de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d\": container with ID starting with de287c58ebe25b242a843fcad3c7d9fb967bb216fb4bdc344091be3189635e5d not found: ID does not exist" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.367036 4788 scope.go:117] "RemoveContainer" containerID="ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.367451 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40"} err="failed to get container status \"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40\": rpc error: code = NotFound desc = could not find container \"ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40\": container with ID starting with ecdfd07f99b8b2d768b1a15a7c7dda68c86e72db99550a9ef8a7a4ac75561f40 not found: ID does not exist" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.610091 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.622678 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.640176 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:33 crc kubenswrapper[4788]: E1010 15:07:33.640602 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62ef92f-3981-4eca-b119-faeb979af999" containerName="nova-metadata-metadata" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.640618 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62ef92f-3981-4eca-b119-faeb979af999" containerName="nova-metadata-metadata" Oct 10 15:07:33 crc kubenswrapper[4788]: E1010 15:07:33.640637 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62ef92f-3981-4eca-b119-faeb979af999" containerName="nova-metadata-log" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.640644 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62ef92f-3981-4eca-b119-faeb979af999" containerName="nova-metadata-log" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.640859 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c62ef92f-3981-4eca-b119-faeb979af999" containerName="nova-metadata-metadata" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.640884 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c62ef92f-3981-4eca-b119-faeb979af999" containerName="nova-metadata-log" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.641954 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.651602 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.652675 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.653542 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.745243 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.745305 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftg6r\" (UniqueName: \"kubernetes.io/projected/de2fb68e-d123-4d2e-8bc2-abae919fb22e-kube-api-access-ftg6r\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.745392 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-config-data\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.745419 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.745437 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2fb68e-d123-4d2e-8bc2-abae919fb22e-logs\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.847678 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-config-data\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.847727 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.847748 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2fb68e-d123-4d2e-8bc2-abae919fb22e-logs\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.847853 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.847870 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftg6r\" (UniqueName: \"kubernetes.io/projected/de2fb68e-d123-4d2e-8bc2-abae919fb22e-kube-api-access-ftg6r\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.848366 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2fb68e-d123-4d2e-8bc2-abae919fb22e-logs\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.857348 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.861303 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.865765 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftg6r\" (UniqueName: \"kubernetes.io/projected/de2fb68e-d123-4d2e-8bc2-abae919fb22e-kube-api-access-ftg6r\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.871776 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-config-data\") pod \"nova-metadata-0\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " pod="openstack/nova-metadata-0" Oct 10 15:07:33 crc kubenswrapper[4788]: I1010 15:07:33.957482 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:07:34 crc kubenswrapper[4788]: I1010 15:07:34.252210 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f684fc3-b2c2-454a-b70e-5a19f9520710" path="/var/lib/kubelet/pods/8f684fc3-b2c2-454a-b70e-5a19f9520710/volumes" Oct 10 15:07:34 crc kubenswrapper[4788]: I1010 15:07:34.253408 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c62ef92f-3981-4eca-b119-faeb979af999" path="/var/lib/kubelet/pods/c62ef92f-3981-4eca-b119-faeb979af999/volumes" Oct 10 15:07:34 crc kubenswrapper[4788]: I1010 15:07:34.287168 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"02360792-0295-42be-a94c-de0e7d2aa376","Type":"ContainerStarted","Data":"e0ce65ec0009a40c1eb733283f7b65324c71d464e88882462d80a65bd07ae1af"} Oct 10 15:07:34 crc kubenswrapper[4788]: I1010 15:07:34.287229 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"02360792-0295-42be-a94c-de0e7d2aa376","Type":"ContainerStarted","Data":"9f77918f6dbccdb5b5e3ebb8be7846a782d1131d10ba2ad042638d4259dee7bf"} Oct 10 15:07:34 crc kubenswrapper[4788]: I1010 15:07:34.287311 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:34 crc kubenswrapper[4788]: I1010 15:07:34.313135 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.313083195 podStartE2EDuration="2.313083195s" podCreationTimestamp="2025-10-10 15:07:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:34.309724884 +0000 UTC m=+1356.759440442" watchObservedRunningTime="2025-10-10 15:07:34.313083195 +0000 UTC m=+1356.762798763" Oct 10 15:07:34 crc kubenswrapper[4788]: I1010 15:07:34.456518 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:07:35 crc kubenswrapper[4788]: I1010 15:07:35.328627 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"de2fb68e-d123-4d2e-8bc2-abae919fb22e","Type":"ContainerStarted","Data":"65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c"} Oct 10 15:07:35 crc kubenswrapper[4788]: I1010 15:07:35.328953 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"de2fb68e-d123-4d2e-8bc2-abae919fb22e","Type":"ContainerStarted","Data":"9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999"} Oct 10 15:07:35 crc kubenswrapper[4788]: I1010 15:07:35.328964 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"de2fb68e-d123-4d2e-8bc2-abae919fb22e","Type":"ContainerStarted","Data":"f87cb49e093f9773649f15564dabd4a93c1852334703cef80c55c84c20857729"} Oct 10 15:07:35 crc kubenswrapper[4788]: I1010 15:07:35.351852 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.351807262 podStartE2EDuration="2.351807262s" podCreationTimestamp="2025-10-10 15:07:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:35.347725851 +0000 UTC m=+1357.797441399" watchObservedRunningTime="2025-10-10 15:07:35.351807262 +0000 UTC m=+1357.801522810" Oct 10 15:07:35 crc kubenswrapper[4788]: E1010 15:07:35.641545 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="155460bd9f7f7a8b06685eb2550e5f97ae4172a9cd51c3ea8ed6a80fa5b5287e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 15:07:35 crc kubenswrapper[4788]: E1010 15:07:35.643528 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="155460bd9f7f7a8b06685eb2550e5f97ae4172a9cd51c3ea8ed6a80fa5b5287e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 15:07:35 crc kubenswrapper[4788]: E1010 15:07:35.647121 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="155460bd9f7f7a8b06685eb2550e5f97ae4172a9cd51c3ea8ed6a80fa5b5287e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 15:07:35 crc kubenswrapper[4788]: E1010 15:07:35.647180 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f3f7a671-e17f-4be0-8afb-d10395da7ad8" containerName="nova-scheduler-scheduler" Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.350522 4788 generic.go:334] "Generic (PLEG): container finished" podID="f3f7a671-e17f-4be0-8afb-d10395da7ad8" containerID="155460bd9f7f7a8b06685eb2550e5f97ae4172a9cd51c3ea8ed6a80fa5b5287e" exitCode=0 Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.350619 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f3f7a671-e17f-4be0-8afb-d10395da7ad8","Type":"ContainerDied","Data":"155460bd9f7f7a8b06685eb2550e5f97ae4172a9cd51c3ea8ed6a80fa5b5287e"} Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.350828 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f3f7a671-e17f-4be0-8afb-d10395da7ad8","Type":"ContainerDied","Data":"67f1bdd906e2aa86153e2ca8c12bce6176cc913f19d7d1bbf378a6ec716f4dd0"} Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.350854 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67f1bdd906e2aa86153e2ca8c12bce6176cc913f19d7d1bbf378a6ec716f4dd0" Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.375857 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.532722 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnxtj\" (UniqueName: \"kubernetes.io/projected/f3f7a671-e17f-4be0-8afb-d10395da7ad8-kube-api-access-fnxtj\") pod \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.532946 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-config-data\") pod \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.533036 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-combined-ca-bundle\") pod \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\" (UID: \"f3f7a671-e17f-4be0-8afb-d10395da7ad8\") " Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.542451 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3f7a671-e17f-4be0-8afb-d10395da7ad8-kube-api-access-fnxtj" (OuterVolumeSpecName: "kube-api-access-fnxtj") pod "f3f7a671-e17f-4be0-8afb-d10395da7ad8" (UID: "f3f7a671-e17f-4be0-8afb-d10395da7ad8"). InnerVolumeSpecName "kube-api-access-fnxtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.578999 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-config-data" (OuterVolumeSpecName: "config-data") pod "f3f7a671-e17f-4be0-8afb-d10395da7ad8" (UID: "f3f7a671-e17f-4be0-8afb-d10395da7ad8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.590479 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3f7a671-e17f-4be0-8afb-d10395da7ad8" (UID: "f3f7a671-e17f-4be0-8afb-d10395da7ad8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.635912 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnxtj\" (UniqueName: \"kubernetes.io/projected/f3f7a671-e17f-4be0-8afb-d10395da7ad8-kube-api-access-fnxtj\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.635950 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:37 crc kubenswrapper[4788]: I1010 15:07:37.635960 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3f7a671-e17f-4be0-8afb-d10395da7ad8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.363070 4788 generic.go:334] "Generic (PLEG): container finished" podID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerID="4104f4cfee1edb339a3c5824fe82691a528e1cc74ee8a2ec60a5c133c43e91ee" exitCode=0 Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.363516 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.364026 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c18dca-feb6-4006-a0ec-817e01a156c9","Type":"ContainerDied","Data":"4104f4cfee1edb339a3c5824fe82691a528e1cc74ee8a2ec60a5c133c43e91ee"} Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.364053 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"78c18dca-feb6-4006-a0ec-817e01a156c9","Type":"ContainerDied","Data":"d6e0994090b1b521e2bfa84d683e5b56386f29fe2b62f7256841b359b7fd4a95"} Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.364066 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6e0994090b1b521e2bfa84d683e5b56386f29fe2b62f7256841b359b7fd4a95" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.388523 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.404273 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.413699 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.458609 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:38 crc kubenswrapper[4788]: E1010 15:07:38.459161 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-log" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.459187 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-log" Oct 10 15:07:38 crc kubenswrapper[4788]: E1010 15:07:38.459229 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-api" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.459238 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-api" Oct 10 15:07:38 crc kubenswrapper[4788]: E1010 15:07:38.459288 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3f7a671-e17f-4be0-8afb-d10395da7ad8" containerName="nova-scheduler-scheduler" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.459297 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3f7a671-e17f-4be0-8afb-d10395da7ad8" containerName="nova-scheduler-scheduler" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.459575 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-api" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.459601 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" containerName="nova-api-log" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.459629 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3f7a671-e17f-4be0-8afb-d10395da7ad8" containerName="nova-scheduler-scheduler" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.460449 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.463321 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.469931 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.564849 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c9wt\" (UniqueName: \"kubernetes.io/projected/78c18dca-feb6-4006-a0ec-817e01a156c9-kube-api-access-7c9wt\") pod \"78c18dca-feb6-4006-a0ec-817e01a156c9\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.565007 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c18dca-feb6-4006-a0ec-817e01a156c9-logs\") pod \"78c18dca-feb6-4006-a0ec-817e01a156c9\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.565299 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-combined-ca-bundle\") pod \"78c18dca-feb6-4006-a0ec-817e01a156c9\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.565506 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-config-data\") pod \"78c18dca-feb6-4006-a0ec-817e01a156c9\" (UID: \"78c18dca-feb6-4006-a0ec-817e01a156c9\") " Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.565615 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78c18dca-feb6-4006-a0ec-817e01a156c9-logs" (OuterVolumeSpecName: "logs") pod "78c18dca-feb6-4006-a0ec-817e01a156c9" (UID: "78c18dca-feb6-4006-a0ec-817e01a156c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.565978 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-config-data\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.566035 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n766m\" (UniqueName: \"kubernetes.io/projected/2da0778e-f2cc-4897-9996-a6973c974954-kube-api-access-n766m\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.566070 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.566359 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78c18dca-feb6-4006-a0ec-817e01a156c9-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.570619 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c18dca-feb6-4006-a0ec-817e01a156c9-kube-api-access-7c9wt" (OuterVolumeSpecName: "kube-api-access-7c9wt") pod "78c18dca-feb6-4006-a0ec-817e01a156c9" (UID: "78c18dca-feb6-4006-a0ec-817e01a156c9"). InnerVolumeSpecName "kube-api-access-7c9wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.594779 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-config-data" (OuterVolumeSpecName: "config-data") pod "78c18dca-feb6-4006-a0ec-817e01a156c9" (UID: "78c18dca-feb6-4006-a0ec-817e01a156c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.595981 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78c18dca-feb6-4006-a0ec-817e01a156c9" (UID: "78c18dca-feb6-4006-a0ec-817e01a156c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.669113 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-config-data\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.669237 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n766m\" (UniqueName: \"kubernetes.io/projected/2da0778e-f2cc-4897-9996-a6973c974954-kube-api-access-n766m\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.669277 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.669895 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c9wt\" (UniqueName: \"kubernetes.io/projected/78c18dca-feb6-4006-a0ec-817e01a156c9-kube-api-access-7c9wt\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.669916 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.669927 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c18dca-feb6-4006-a0ec-817e01a156c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.675919 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.676254 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-config-data\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.689942 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n766m\" (UniqueName: \"kubernetes.io/projected/2da0778e-f2cc-4897-9996-a6973c974954-kube-api-access-n766m\") pod \"nova-scheduler-0\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.790990 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.958354 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 15:07:38 crc kubenswrapper[4788]: I1010 15:07:38.958899 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.116945 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.269606 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.373055 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.383741 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2da0778e-f2cc-4897-9996-a6973c974954","Type":"ContainerStarted","Data":"31dfedc38abb192f7849761f63090ac4313d210060076fdee2eb412af7d5f48a"} Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.458660 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.467678 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.490908 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.493416 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.497023 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.504438 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.591124 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.591690 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-config-data\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.591723 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-logs\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.591769 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgqq2\" (UniqueName: \"kubernetes.io/projected/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-kube-api-access-vgqq2\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.693877 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.693962 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-config-data\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.694007 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-logs\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.694068 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgqq2\" (UniqueName: \"kubernetes.io/projected/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-kube-api-access-vgqq2\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.694592 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-logs\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.700839 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.701558 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-config-data\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.715601 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgqq2\" (UniqueName: \"kubernetes.io/projected/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-kube-api-access-vgqq2\") pod \"nova-api-0\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " pod="openstack/nova-api-0" Oct 10 15:07:39 crc kubenswrapper[4788]: I1010 15:07:39.815241 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:07:40 crc kubenswrapper[4788]: I1010 15:07:40.248841 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c18dca-feb6-4006-a0ec-817e01a156c9" path="/var/lib/kubelet/pods/78c18dca-feb6-4006-a0ec-817e01a156c9/volumes" Oct 10 15:07:40 crc kubenswrapper[4788]: I1010 15:07:40.249793 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3f7a671-e17f-4be0-8afb-d10395da7ad8" path="/var/lib/kubelet/pods/f3f7a671-e17f-4be0-8afb-d10395da7ad8/volumes" Oct 10 15:07:40 crc kubenswrapper[4788]: I1010 15:07:40.348769 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:07:40 crc kubenswrapper[4788]: W1010 15:07:40.358931 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda57f5529_f3b7_4620_b81a_4ab31c3a5dda.slice/crio-03b5831ddd3efe4c3e0de7f1418e5d01a12522d00569c19c23c3dc0c356907d7 WatchSource:0}: Error finding container 03b5831ddd3efe4c3e0de7f1418e5d01a12522d00569c19c23c3dc0c356907d7: Status 404 returned error can't find the container with id 03b5831ddd3efe4c3e0de7f1418e5d01a12522d00569c19c23c3dc0c356907d7 Oct 10 15:07:40 crc kubenswrapper[4788]: I1010 15:07:40.392851 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a57f5529-f3b7-4620-b81a-4ab31c3a5dda","Type":"ContainerStarted","Data":"03b5831ddd3efe4c3e0de7f1418e5d01a12522d00569c19c23c3dc0c356907d7"} Oct 10 15:07:40 crc kubenswrapper[4788]: I1010 15:07:40.394889 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2da0778e-f2cc-4897-9996-a6973c974954","Type":"ContainerStarted","Data":"50b6d48d5eeced1e1b4d152b6f21ebd20a78a742d49a57f9611f4997f62ab50e"} Oct 10 15:07:40 crc kubenswrapper[4788]: I1010 15:07:40.420591 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.4205619 podStartE2EDuration="2.4205619s" podCreationTimestamp="2025-10-10 15:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:40.414851526 +0000 UTC m=+1362.864567074" watchObservedRunningTime="2025-10-10 15:07:40.4205619 +0000 UTC m=+1362.870277458" Oct 10 15:07:41 crc kubenswrapper[4788]: I1010 15:07:41.413478 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a57f5529-f3b7-4620-b81a-4ab31c3a5dda","Type":"ContainerStarted","Data":"825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e"} Oct 10 15:07:41 crc kubenswrapper[4788]: I1010 15:07:41.414360 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a57f5529-f3b7-4620-b81a-4ab31c3a5dda","Type":"ContainerStarted","Data":"2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7"} Oct 10 15:07:41 crc kubenswrapper[4788]: I1010 15:07:41.457861 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.457824951 podStartE2EDuration="2.457824951s" podCreationTimestamp="2025-10-10 15:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:41.441566602 +0000 UTC m=+1363.891282160" watchObservedRunningTime="2025-10-10 15:07:41.457824951 +0000 UTC m=+1363.907540509" Oct 10 15:07:42 crc kubenswrapper[4788]: I1010 15:07:42.676408 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 10 15:07:42 crc kubenswrapper[4788]: I1010 15:07:42.922562 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:07:42 crc kubenswrapper[4788]: I1010 15:07:42.922845 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c" containerName="kube-state-metrics" containerID="cri-o://6dfca80283e228ab25b825ce8120c260fb7c568dd80ba2c02f1a941cc12700a6" gracePeriod=30 Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.437644 4788 generic.go:334] "Generic (PLEG): container finished" podID="23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c" containerID="6dfca80283e228ab25b825ce8120c260fb7c568dd80ba2c02f1a941cc12700a6" exitCode=2 Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.437971 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c","Type":"ContainerDied","Data":"6dfca80283e228ab25b825ce8120c260fb7c568dd80ba2c02f1a941cc12700a6"} Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.438001 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c","Type":"ContainerDied","Data":"e1fb1fc56de484f2a92aeafcc986d51404f6a3b8e03753d1b05a2d8adf62ce76"} Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.438011 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1fb1fc56de484f2a92aeafcc986d51404f6a3b8e03753d1b05a2d8adf62ce76" Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.475609 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.582478 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trkbn\" (UniqueName: \"kubernetes.io/projected/23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c-kube-api-access-trkbn\") pod \"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c\" (UID: \"23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c\") " Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.598899 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c-kube-api-access-trkbn" (OuterVolumeSpecName: "kube-api-access-trkbn") pod "23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c" (UID: "23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c"). InnerVolumeSpecName "kube-api-access-trkbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.684867 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trkbn\" (UniqueName: \"kubernetes.io/projected/23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c-kube-api-access-trkbn\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.792887 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.957680 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 15:07:43 crc kubenswrapper[4788]: I1010 15:07:43.958037 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.450234 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.500243 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.512170 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.522470 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:07:44 crc kubenswrapper[4788]: E1010 15:07:44.523418 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c" containerName="kube-state-metrics" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.523524 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c" containerName="kube-state-metrics" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.523897 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c" containerName="kube-state-metrics" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.525052 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.528443 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.528746 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.530391 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.706488 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.706940 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kml9j\" (UniqueName: \"kubernetes.io/projected/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-api-access-kml9j\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.707027 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.707194 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.809233 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kml9j\" (UniqueName: \"kubernetes.io/projected/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-api-access-kml9j\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.809309 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.809353 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.809418 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.816891 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.817099 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.819839 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.847949 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kml9j\" (UniqueName: \"kubernetes.io/projected/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-api-access-kml9j\") pod \"kube-state-metrics-0\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.852986 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.853443 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="proxy-httpd" containerID="cri-o://11b44c1f6365a782b0b7cb7559f79e19ab5b11e060b32f8b117349ae62639bc0" gracePeriod=30 Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.853588 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="ceilometer-notification-agent" containerID="cri-o://4c3acbecf325173d28f79d929e2c02d58e460088b059fe2cd30d126c41591ab3" gracePeriod=30 Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.853593 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="ceilometer-central-agent" containerID="cri-o://2d6153d79b9a0c23fdfaba336c12b1d4239ff99581357a3dddaa2a13e66e9499" gracePeriod=30 Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.853696 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="sg-core" containerID="cri-o://3d5409a1cf47943139f8aadab679681c2155735e3b31ea44ebf9c3f8a684c34a" gracePeriod=30 Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.854087 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.979524 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 15:07:44 crc kubenswrapper[4788]: I1010 15:07:44.979896 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 15:07:45 crc kubenswrapper[4788]: I1010 15:07:45.451873 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:07:45 crc kubenswrapper[4788]: I1010 15:07:45.465334 4788 generic.go:334] "Generic (PLEG): container finished" podID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerID="11b44c1f6365a782b0b7cb7559f79e19ab5b11e060b32f8b117349ae62639bc0" exitCode=0 Oct 10 15:07:45 crc kubenswrapper[4788]: I1010 15:07:45.465374 4788 generic.go:334] "Generic (PLEG): container finished" podID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerID="3d5409a1cf47943139f8aadab679681c2155735e3b31ea44ebf9c3f8a684c34a" exitCode=2 Oct 10 15:07:45 crc kubenswrapper[4788]: I1010 15:07:45.465383 4788 generic.go:334] "Generic (PLEG): container finished" podID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerID="2d6153d79b9a0c23fdfaba336c12b1d4239ff99581357a3dddaa2a13e66e9499" exitCode=0 Oct 10 15:07:45 crc kubenswrapper[4788]: I1010 15:07:45.465454 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerDied","Data":"11b44c1f6365a782b0b7cb7559f79e19ab5b11e060b32f8b117349ae62639bc0"} Oct 10 15:07:45 crc kubenswrapper[4788]: I1010 15:07:45.465486 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerDied","Data":"3d5409a1cf47943139f8aadab679681c2155735e3b31ea44ebf9c3f8a684c34a"} Oct 10 15:07:45 crc kubenswrapper[4788]: I1010 15:07:45.465496 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerDied","Data":"2d6153d79b9a0c23fdfaba336c12b1d4239ff99581357a3dddaa2a13e66e9499"} Oct 10 15:07:45 crc kubenswrapper[4788]: I1010 15:07:45.466422 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f13a166-2b9b-4679-877b-f4596b2d45ed","Type":"ContainerStarted","Data":"93b5237c52064c0a31349ac3298588c7ce002ea184a528a2127e6fc9caff6e44"} Oct 10 15:07:46 crc kubenswrapper[4788]: I1010 15:07:46.245648 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c" path="/var/lib/kubelet/pods/23e2bf75-ba7d-4ca8-a9e0-c94b4aec765c/volumes" Oct 10 15:07:46 crc kubenswrapper[4788]: I1010 15:07:46.479283 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f13a166-2b9b-4679-877b-f4596b2d45ed","Type":"ContainerStarted","Data":"22d302e6d9d9eed7daebd5122f717fd594b1258085c469fab8d22ea7ffda2b6e"} Oct 10 15:07:46 crc kubenswrapper[4788]: I1010 15:07:46.479551 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 10 15:07:46 crc kubenswrapper[4788]: I1010 15:07:46.510063 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.154758508 podStartE2EDuration="2.510039992s" podCreationTimestamp="2025-10-10 15:07:44 +0000 UTC" firstStartedPulling="2025-10-10 15:07:45.447291724 +0000 UTC m=+1367.897007272" lastFinishedPulling="2025-10-10 15:07:45.802573208 +0000 UTC m=+1368.252288756" observedRunningTime="2025-10-10 15:07:46.498275495 +0000 UTC m=+1368.947991043" watchObservedRunningTime="2025-10-10 15:07:46.510039992 +0000 UTC m=+1368.959755540" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.501552 4788 generic.go:334] "Generic (PLEG): container finished" podID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerID="4c3acbecf325173d28f79d929e2c02d58e460088b059fe2cd30d126c41591ab3" exitCode=0 Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.501647 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerDied","Data":"4c3acbecf325173d28f79d929e2c02d58e460088b059fe2cd30d126c41591ab3"} Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.639824 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.695159 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-log-httpd\") pod \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.695231 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-combined-ca-bundle\") pod \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.695307 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-sg-core-conf-yaml\") pod \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.695348 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-scripts\") pod \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.695397 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-config-data\") pod \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.695424 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48bkc\" (UniqueName: \"kubernetes.io/projected/feb1c7a6-0737-4d42-8c4e-be384b5a4074-kube-api-access-48bkc\") pod \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.695460 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-run-httpd\") pod \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.695975 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "feb1c7a6-0737-4d42-8c4e-be384b5a4074" (UID: "feb1c7a6-0737-4d42-8c4e-be384b5a4074"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.696488 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "feb1c7a6-0737-4d42-8c4e-be384b5a4074" (UID: "feb1c7a6-0737-4d42-8c4e-be384b5a4074"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.704263 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feb1c7a6-0737-4d42-8c4e-be384b5a4074-kube-api-access-48bkc" (OuterVolumeSpecName: "kube-api-access-48bkc") pod "feb1c7a6-0737-4d42-8c4e-be384b5a4074" (UID: "feb1c7a6-0737-4d42-8c4e-be384b5a4074"). InnerVolumeSpecName "kube-api-access-48bkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.711217 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-scripts" (OuterVolumeSpecName: "scripts") pod "feb1c7a6-0737-4d42-8c4e-be384b5a4074" (UID: "feb1c7a6-0737-4d42-8c4e-be384b5a4074"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.727318 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "feb1c7a6-0737-4d42-8c4e-be384b5a4074" (UID: "feb1c7a6-0737-4d42-8c4e-be384b5a4074"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.796316 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "feb1c7a6-0737-4d42-8c4e-be384b5a4074" (UID: "feb1c7a6-0737-4d42-8c4e-be384b5a4074"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.796833 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-combined-ca-bundle\") pod \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\" (UID: \"feb1c7a6-0737-4d42-8c4e-be384b5a4074\") " Oct 10 15:07:47 crc kubenswrapper[4788]: W1010 15:07:47.797037 4788 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/feb1c7a6-0737-4d42-8c4e-be384b5a4074/volumes/kubernetes.io~secret/combined-ca-bundle Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.797066 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "feb1c7a6-0737-4d42-8c4e-be384b5a4074" (UID: "feb1c7a6-0737-4d42-8c4e-be384b5a4074"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.797269 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.797285 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.797296 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.797306 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.797316 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48bkc\" (UniqueName: \"kubernetes.io/projected/feb1c7a6-0737-4d42-8c4e-be384b5a4074-kube-api-access-48bkc\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.797324 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/feb1c7a6-0737-4d42-8c4e-be384b5a4074-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.816544 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-config-data" (OuterVolumeSpecName: "config-data") pod "feb1c7a6-0737-4d42-8c4e-be384b5a4074" (UID: "feb1c7a6-0737-4d42-8c4e-be384b5a4074"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:47 crc kubenswrapper[4788]: I1010 15:07:47.898684 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feb1c7a6-0737-4d42-8c4e-be384b5a4074-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.515600 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"feb1c7a6-0737-4d42-8c4e-be384b5a4074","Type":"ContainerDied","Data":"a28995f507205de0012a07bbf23fe5bdfc9d314e572c4f976638c6fc1e49da8e"} Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.515684 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.516239 4788 scope.go:117] "RemoveContainer" containerID="11b44c1f6365a782b0b7cb7559f79e19ab5b11e060b32f8b117349ae62639bc0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.540884 4788 scope.go:117] "RemoveContainer" containerID="3d5409a1cf47943139f8aadab679681c2155735e3b31ea44ebf9c3f8a684c34a" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.562713 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.585285 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.606645 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:48 crc kubenswrapper[4788]: E1010 15:07:48.607051 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="sg-core" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.607064 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="sg-core" Oct 10 15:07:48 crc kubenswrapper[4788]: E1010 15:07:48.607078 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="ceilometer-central-agent" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.607087 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="ceilometer-central-agent" Oct 10 15:07:48 crc kubenswrapper[4788]: E1010 15:07:48.607131 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="ceilometer-notification-agent" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.607151 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="ceilometer-notification-agent" Oct 10 15:07:48 crc kubenswrapper[4788]: E1010 15:07:48.607182 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="proxy-httpd" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.607190 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="proxy-httpd" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.607378 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="proxy-httpd" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.607396 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="ceilometer-notification-agent" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.607409 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="ceilometer-central-agent" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.607427 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" containerName="sg-core" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.609112 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.609242 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.618436 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.618538 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.618680 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.621581 4788 scope.go:117] "RemoveContainer" containerID="4c3acbecf325173d28f79d929e2c02d58e460088b059fe2cd30d126c41591ab3" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.659669 4788 scope.go:117] "RemoveContainer" containerID="2d6153d79b9a0c23fdfaba336c12b1d4239ff99581357a3dddaa2a13e66e9499" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.720072 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-scripts\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.720214 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.720330 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.720752 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.720884 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-config-data\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.720985 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-log-httpd\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.721043 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltm7r\" (UniqueName: \"kubernetes.io/projected/d856f37d-ac36-4748-974c-f4ad13b594a5-kube-api-access-ltm7r\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.721402 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-run-httpd\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.792519 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.818928 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.824959 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-run-httpd\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825025 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-scripts\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825195 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825215 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825240 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825276 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-config-data\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825319 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-log-httpd\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825355 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltm7r\" (UniqueName: \"kubernetes.io/projected/d856f37d-ac36-4748-974c-f4ad13b594a5-kube-api-access-ltm7r\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825611 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-run-httpd\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.825953 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-log-httpd\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.830256 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.831917 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.836493 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.837368 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-scripts\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.840391 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-config-data\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.841466 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltm7r\" (UniqueName: \"kubernetes.io/projected/d856f37d-ac36-4748-974c-f4ad13b594a5-kube-api-access-ltm7r\") pod \"ceilometer-0\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " pod="openstack/ceilometer-0" Oct 10 15:07:48 crc kubenswrapper[4788]: I1010 15:07:48.953232 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:07:49 crc kubenswrapper[4788]: I1010 15:07:49.418907 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:07:49 crc kubenswrapper[4788]: W1010 15:07:49.422947 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd856f37d_ac36_4748_974c_f4ad13b594a5.slice/crio-403b5d802ef4b3aef4099d634b839eff2f7e7422536a0b36ada0b83105ad497f WatchSource:0}: Error finding container 403b5d802ef4b3aef4099d634b839eff2f7e7422536a0b36ada0b83105ad497f: Status 404 returned error can't find the container with id 403b5d802ef4b3aef4099d634b839eff2f7e7422536a0b36ada0b83105ad497f Oct 10 15:07:49 crc kubenswrapper[4788]: I1010 15:07:49.528267 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerStarted","Data":"403b5d802ef4b3aef4099d634b839eff2f7e7422536a0b36ada0b83105ad497f"} Oct 10 15:07:49 crc kubenswrapper[4788]: I1010 15:07:49.564850 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 15:07:49 crc kubenswrapper[4788]: I1010 15:07:49.817125 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 15:07:49 crc kubenswrapper[4788]: I1010 15:07:49.817710 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 15:07:50 crc kubenswrapper[4788]: I1010 15:07:50.251466 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feb1c7a6-0737-4d42-8c4e-be384b5a4074" path="/var/lib/kubelet/pods/feb1c7a6-0737-4d42-8c4e-be384b5a4074/volumes" Oct 10 15:07:50 crc kubenswrapper[4788]: I1010 15:07:50.564892 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerStarted","Data":"d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25"} Oct 10 15:07:50 crc kubenswrapper[4788]: I1010 15:07:50.900590 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 15:07:50 crc kubenswrapper[4788]: I1010 15:07:50.900899 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 15:07:51 crc kubenswrapper[4788]: I1010 15:07:51.608918 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerStarted","Data":"89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239"} Oct 10 15:07:52 crc kubenswrapper[4788]: I1010 15:07:52.618943 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerStarted","Data":"7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e"} Oct 10 15:07:53 crc kubenswrapper[4788]: I1010 15:07:53.632010 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerStarted","Data":"75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c"} Oct 10 15:07:53 crc kubenswrapper[4788]: I1010 15:07:53.633316 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 15:07:53 crc kubenswrapper[4788]: I1010 15:07:53.657468 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.324557092 podStartE2EDuration="5.657447103s" podCreationTimestamp="2025-10-10 15:07:48 +0000 UTC" firstStartedPulling="2025-10-10 15:07:49.42569893 +0000 UTC m=+1371.875414488" lastFinishedPulling="2025-10-10 15:07:52.758588951 +0000 UTC m=+1375.208304499" observedRunningTime="2025-10-10 15:07:53.651531894 +0000 UTC m=+1376.101247452" watchObservedRunningTime="2025-10-10 15:07:53.657447103 +0000 UTC m=+1376.107162651" Oct 10 15:07:53 crc kubenswrapper[4788]: I1010 15:07:53.963193 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 15:07:53 crc kubenswrapper[4788]: I1010 15:07:53.964857 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 15:07:53 crc kubenswrapper[4788]: I1010 15:07:53.971544 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 15:07:54 crc kubenswrapper[4788]: I1010 15:07:54.647489 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 15:07:54 crc kubenswrapper[4788]: I1010 15:07:54.878622 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.436536 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.572525 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7xbt\" (UniqueName: \"kubernetes.io/projected/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-kube-api-access-k7xbt\") pod \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.572640 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-combined-ca-bundle\") pod \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.572827 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-config-data\") pod \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\" (UID: \"3f0bdf94-e96f-47ca-96a6-677a9546fc4d\") " Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.578823 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-kube-api-access-k7xbt" (OuterVolumeSpecName: "kube-api-access-k7xbt") pod "3f0bdf94-e96f-47ca-96a6-677a9546fc4d" (UID: "3f0bdf94-e96f-47ca-96a6-677a9546fc4d"). InnerVolumeSpecName "kube-api-access-k7xbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.602263 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-config-data" (OuterVolumeSpecName: "config-data") pod "3f0bdf94-e96f-47ca-96a6-677a9546fc4d" (UID: "3f0bdf94-e96f-47ca-96a6-677a9546fc4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.602628 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f0bdf94-e96f-47ca-96a6-677a9546fc4d" (UID: "3f0bdf94-e96f-47ca-96a6-677a9546fc4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.649717 4788 generic.go:334] "Generic (PLEG): container finished" podID="3f0bdf94-e96f-47ca-96a6-677a9546fc4d" containerID="226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb" exitCode=137 Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.649808 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.650985 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f0bdf94-e96f-47ca-96a6-677a9546fc4d","Type":"ContainerDied","Data":"226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb"} Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.652181 4788 scope.go:117] "RemoveContainer" containerID="226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.655825 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3f0bdf94-e96f-47ca-96a6-677a9546fc4d","Type":"ContainerDied","Data":"3d17c2ebcd07aad38bba0af6889885d35ea62ce5db7407a74accad7dec18861c"} Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.678361 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.678442 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7xbt\" (UniqueName: \"kubernetes.io/projected/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-kube-api-access-k7xbt\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.678490 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f0bdf94-e96f-47ca-96a6-677a9546fc4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.696399 4788 scope.go:117] "RemoveContainer" containerID="226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb" Oct 10 15:07:55 crc kubenswrapper[4788]: E1010 15:07:55.696964 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb\": container with ID starting with 226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb not found: ID does not exist" containerID="226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.697021 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb"} err="failed to get container status \"226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb\": rpc error: code = NotFound desc = could not find container \"226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb\": container with ID starting with 226e4d789ea4fd5930b39805280581560d1aae3aa6b91cc97599e9ca1bf9cffb not found: ID does not exist" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.705473 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.721679 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.733579 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:55 crc kubenswrapper[4788]: E1010 15:07:55.734053 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f0bdf94-e96f-47ca-96a6-677a9546fc4d" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.734083 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f0bdf94-e96f-47ca-96a6-677a9546fc4d" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.734371 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f0bdf94-e96f-47ca-96a6-677a9546fc4d" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.736289 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.742262 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.742608 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.743026 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.758449 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.882657 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.883978 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjfvm\" (UniqueName: \"kubernetes.io/projected/c35da072-521f-4266-9600-688bf5b49c1c-kube-api-access-pjfvm\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.884205 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.884361 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.884407 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.986661 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjfvm\" (UniqueName: \"kubernetes.io/projected/c35da072-521f-4266-9600-688bf5b49c1c-kube-api-access-pjfvm\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.986783 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.986829 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.986851 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.986880 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.990552 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.990727 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.990820 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:55 crc kubenswrapper[4788]: I1010 15:07:55.991382 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:56 crc kubenswrapper[4788]: I1010 15:07:56.003510 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjfvm\" (UniqueName: \"kubernetes.io/projected/c35da072-521f-4266-9600-688bf5b49c1c-kube-api-access-pjfvm\") pod \"nova-cell1-novncproxy-0\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:56 crc kubenswrapper[4788]: I1010 15:07:56.054967 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:07:56 crc kubenswrapper[4788]: I1010 15:07:56.245969 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f0bdf94-e96f-47ca-96a6-677a9546fc4d" path="/var/lib/kubelet/pods/3f0bdf94-e96f-47ca-96a6-677a9546fc4d/volumes" Oct 10 15:07:56 crc kubenswrapper[4788]: I1010 15:07:56.581959 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:07:56 crc kubenswrapper[4788]: I1010 15:07:56.664557 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c35da072-521f-4266-9600-688bf5b49c1c","Type":"ContainerStarted","Data":"bf9f470005274a391d3390ce8ec6e5159f552cd80b56c76a008259e229d503f8"} Oct 10 15:07:57 crc kubenswrapper[4788]: I1010 15:07:57.685675 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c35da072-521f-4266-9600-688bf5b49c1c","Type":"ContainerStarted","Data":"d0bc3210873389e2972e8761b9e10f4be661565d640a62e62ffc9f730bc70876"} Oct 10 15:07:57 crc kubenswrapper[4788]: I1010 15:07:57.707635 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.707600631 podStartE2EDuration="2.707600631s" podCreationTimestamp="2025-10-10 15:07:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:07:57.70459546 +0000 UTC m=+1380.154310998" watchObservedRunningTime="2025-10-10 15:07:57.707600631 +0000 UTC m=+1380.157316209" Oct 10 15:07:59 crc kubenswrapper[4788]: I1010 15:07:59.822526 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 15:07:59 crc kubenswrapper[4788]: I1010 15:07:59.823772 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 15:07:59 crc kubenswrapper[4788]: I1010 15:07:59.829134 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 15:07:59 crc kubenswrapper[4788]: I1010 15:07:59.829310 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 15:08:00 crc kubenswrapper[4788]: I1010 15:08:00.719549 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 15:08:00 crc kubenswrapper[4788]: I1010 15:08:00.725072 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 15:08:00 crc kubenswrapper[4788]: I1010 15:08:00.934198 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wkjwc"] Oct 10 15:08:00 crc kubenswrapper[4788]: I1010 15:08:00.936361 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:00 crc kubenswrapper[4788]: I1010 15:08:00.957399 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wkjwc"] Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.055301 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.133274 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrx79\" (UniqueName: \"kubernetes.io/projected/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-kube-api-access-xrx79\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.133326 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.133395 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.133424 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.133446 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.133487 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-config\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.235826 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrx79\" (UniqueName: \"kubernetes.io/projected/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-kube-api-access-xrx79\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.235880 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.236006 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.236069 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.236112 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.236215 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-config\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.237727 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.237731 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-config\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.238104 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.238819 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.238840 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.269544 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrx79\" (UniqueName: \"kubernetes.io/projected/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-kube-api-access-xrx79\") pod \"dnsmasq-dns-59cf4bdb65-wkjwc\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:01 crc kubenswrapper[4788]: I1010 15:08:01.564816 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:02 crc kubenswrapper[4788]: I1010 15:08:02.084414 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wkjwc"] Oct 10 15:08:02 crc kubenswrapper[4788]: I1010 15:08:02.751803 4788 generic.go:334] "Generic (PLEG): container finished" podID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" containerID="70d6c0ca57d575e5a4f3a8e77dbaf0ffac929fea441a817e22ab138834ce2921" exitCode=0 Oct 10 15:08:02 crc kubenswrapper[4788]: I1010 15:08:02.752045 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" event={"ID":"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed","Type":"ContainerDied","Data":"70d6c0ca57d575e5a4f3a8e77dbaf0ffac929fea441a817e22ab138834ce2921"} Oct 10 15:08:02 crc kubenswrapper[4788]: I1010 15:08:02.752658 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" event={"ID":"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed","Type":"ContainerStarted","Data":"4efcd16f9cd65426df3b6309b3106a3ae25cf253d498d57c6f8be9e8fdda3e1c"} Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.073407 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.074116 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="ceilometer-central-agent" containerID="cri-o://d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25" gracePeriod=30 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.074250 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="proxy-httpd" containerID="cri-o://75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c" gracePeriod=30 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.074313 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="sg-core" containerID="cri-o://7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e" gracePeriod=30 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.074366 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="ceilometer-notification-agent" containerID="cri-o://89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239" gracePeriod=30 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.082329 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.196:3000/\": read tcp 10.217.0.2:52558->10.217.0.196:3000: read: connection reset by peer" Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.189209 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.766370 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" event={"ID":"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed","Type":"ContainerStarted","Data":"a671f2494a1de22a55e001fea5624da5a53f37afb7bb091aee9590f2b7f67b6b"} Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.766431 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.772905 4788 generic.go:334] "Generic (PLEG): container finished" podID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerID="75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c" exitCode=0 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.772948 4788 generic.go:334] "Generic (PLEG): container finished" podID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerID="7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e" exitCode=2 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.772957 4788 generic.go:334] "Generic (PLEG): container finished" podID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerID="d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25" exitCode=0 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.773216 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-log" containerID="cri-o://2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7" gracePeriod=30 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.773315 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-api" containerID="cri-o://825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e" gracePeriod=30 Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.773322 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerDied","Data":"75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c"} Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.773470 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerDied","Data":"7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e"} Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.773496 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerDied","Data":"d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25"} Oct 10 15:08:03 crc kubenswrapper[4788]: I1010 15:08:03.792416 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" podStartSLOduration=3.792393598 podStartE2EDuration="3.792393598s" podCreationTimestamp="2025-10-10 15:08:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:08:03.788040511 +0000 UTC m=+1386.237756059" watchObservedRunningTime="2025-10-10 15:08:03.792393598 +0000 UTC m=+1386.242109146" Oct 10 15:08:04 crc kubenswrapper[4788]: I1010 15:08:04.795191 4788 generic.go:334] "Generic (PLEG): container finished" podID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerID="2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7" exitCode=143 Oct 10 15:08:04 crc kubenswrapper[4788]: I1010 15:08:04.795313 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a57f5529-f3b7-4620-b81a-4ab31c3a5dda","Type":"ContainerDied","Data":"2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7"} Oct 10 15:08:06 crc kubenswrapper[4788]: I1010 15:08:06.055501 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:08:06 crc kubenswrapper[4788]: I1010 15:08:06.083413 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:08:06 crc kubenswrapper[4788]: I1010 15:08:06.841694 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.061925 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2sjbr"] Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.073613 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.075797 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.076052 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.077568 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2sjbr"] Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.170443 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.170542 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-config-data\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.170571 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-scripts\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.170601 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc5wc\" (UniqueName: \"kubernetes.io/projected/b0efd45c-b649-47ad-a008-349c8380a27a-kube-api-access-zc5wc\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.273276 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-config-data\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.273337 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-scripts\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.273388 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc5wc\" (UniqueName: \"kubernetes.io/projected/b0efd45c-b649-47ad-a008-349c8380a27a-kube-api-access-zc5wc\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.273590 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.284824 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.285433 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-config-data\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.316895 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-scripts\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.316982 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc5wc\" (UniqueName: \"kubernetes.io/projected/b0efd45c-b649-47ad-a008-349c8380a27a-kube-api-access-zc5wc\") pod \"nova-cell1-cell-mapping-2sjbr\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.409563 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.459904 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.587126 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-config-data\") pod \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.587404 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgqq2\" (UniqueName: \"kubernetes.io/projected/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-kube-api-access-vgqq2\") pod \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.587463 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-logs\") pod \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.587505 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-combined-ca-bundle\") pod \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\" (UID: \"a57f5529-f3b7-4620-b81a-4ab31c3a5dda\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.588628 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-logs" (OuterVolumeSpecName: "logs") pod "a57f5529-f3b7-4620-b81a-4ab31c3a5dda" (UID: "a57f5529-f3b7-4620-b81a-4ab31c3a5dda"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.597269 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-kube-api-access-vgqq2" (OuterVolumeSpecName: "kube-api-access-vgqq2") pod "a57f5529-f3b7-4620-b81a-4ab31c3a5dda" (UID: "a57f5529-f3b7-4620-b81a-4ab31c3a5dda"). InnerVolumeSpecName "kube-api-access-vgqq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.624993 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-config-data" (OuterVolumeSpecName: "config-data") pod "a57f5529-f3b7-4620-b81a-4ab31c3a5dda" (UID: "a57f5529-f3b7-4620-b81a-4ab31c3a5dda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.636287 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a57f5529-f3b7-4620-b81a-4ab31c3a5dda" (UID: "a57f5529-f3b7-4620-b81a-4ab31c3a5dda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.655064 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.689874 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgqq2\" (UniqueName: \"kubernetes.io/projected/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-kube-api-access-vgqq2\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.690436 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.690649 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.690776 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a57f5529-f3b7-4620-b81a-4ab31c3a5dda-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.792317 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-config-data\") pod \"d856f37d-ac36-4748-974c-f4ad13b594a5\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.792398 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-sg-core-conf-yaml\") pod \"d856f37d-ac36-4748-974c-f4ad13b594a5\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.792520 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-ceilometer-tls-certs\") pod \"d856f37d-ac36-4748-974c-f4ad13b594a5\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.792707 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-run-httpd\") pod \"d856f37d-ac36-4748-974c-f4ad13b594a5\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.792905 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-log-httpd\") pod \"d856f37d-ac36-4748-974c-f4ad13b594a5\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.792957 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-scripts\") pod \"d856f37d-ac36-4748-974c-f4ad13b594a5\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.792983 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-combined-ca-bundle\") pod \"d856f37d-ac36-4748-974c-f4ad13b594a5\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.793044 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltm7r\" (UniqueName: \"kubernetes.io/projected/d856f37d-ac36-4748-974c-f4ad13b594a5-kube-api-access-ltm7r\") pod \"d856f37d-ac36-4748-974c-f4ad13b594a5\" (UID: \"d856f37d-ac36-4748-974c-f4ad13b594a5\") " Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.793338 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d856f37d-ac36-4748-974c-f4ad13b594a5" (UID: "d856f37d-ac36-4748-974c-f4ad13b594a5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.793752 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.793763 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d856f37d-ac36-4748-974c-f4ad13b594a5" (UID: "d856f37d-ac36-4748-974c-f4ad13b594a5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.797903 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-scripts" (OuterVolumeSpecName: "scripts") pod "d856f37d-ac36-4748-974c-f4ad13b594a5" (UID: "d856f37d-ac36-4748-974c-f4ad13b594a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.798579 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d856f37d-ac36-4748-974c-f4ad13b594a5-kube-api-access-ltm7r" (OuterVolumeSpecName: "kube-api-access-ltm7r") pod "d856f37d-ac36-4748-974c-f4ad13b594a5" (UID: "d856f37d-ac36-4748-974c-f4ad13b594a5"). InnerVolumeSpecName "kube-api-access-ltm7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.845415 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d856f37d-ac36-4748-974c-f4ad13b594a5" (UID: "d856f37d-ac36-4748-974c-f4ad13b594a5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.850474 4788 generic.go:334] "Generic (PLEG): container finished" podID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerID="89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239" exitCode=0 Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.850819 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.850875 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerDied","Data":"89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239"} Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.850941 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d856f37d-ac36-4748-974c-f4ad13b594a5","Type":"ContainerDied","Data":"403b5d802ef4b3aef4099d634b839eff2f7e7422536a0b36ada0b83105ad497f"} Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.850969 4788 scope.go:117] "RemoveContainer" containerID="75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.858315 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d856f37d-ac36-4748-974c-f4ad13b594a5" (UID: "d856f37d-ac36-4748-974c-f4ad13b594a5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.859995 4788 generic.go:334] "Generic (PLEG): container finished" podID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerID="825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e" exitCode=0 Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.861808 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.862002 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a57f5529-f3b7-4620-b81a-4ab31c3a5dda","Type":"ContainerDied","Data":"825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e"} Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.862048 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a57f5529-f3b7-4620-b81a-4ab31c3a5dda","Type":"ContainerDied","Data":"03b5831ddd3efe4c3e0de7f1418e5d01a12522d00569c19c23c3dc0c356907d7"} Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.878659 4788 scope.go:117] "RemoveContainer" containerID="7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.895768 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d856f37d-ac36-4748-974c-f4ad13b594a5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.895802 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.895813 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltm7r\" (UniqueName: \"kubernetes.io/projected/d856f37d-ac36-4748-974c-f4ad13b594a5-kube-api-access-ltm7r\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.895824 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.895833 4788 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.905277 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d856f37d-ac36-4748-974c-f4ad13b594a5" (UID: "d856f37d-ac36-4748-974c-f4ad13b594a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.912203 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.916009 4788 scope.go:117] "RemoveContainer" containerID="89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.934988 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.935567 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-config-data" (OuterVolumeSpecName: "config-data") pod "d856f37d-ac36-4748-974c-f4ad13b594a5" (UID: "d856f37d-ac36-4748-974c-f4ad13b594a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.954656 4788 scope.go:117] "RemoveContainer" containerID="d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.973794 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:07 crc kubenswrapper[4788]: E1010 15:08:07.974886 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="ceilometer-central-agent" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.974912 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="ceilometer-central-agent" Oct 10 15:08:07 crc kubenswrapper[4788]: E1010 15:08:07.974954 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-api" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.974963 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-api" Oct 10 15:08:07 crc kubenswrapper[4788]: E1010 15:08:07.974979 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="sg-core" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.974985 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="sg-core" Oct 10 15:08:07 crc kubenswrapper[4788]: E1010 15:08:07.975005 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-log" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975011 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-log" Oct 10 15:08:07 crc kubenswrapper[4788]: E1010 15:08:07.975032 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="ceilometer-notification-agent" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975038 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="ceilometer-notification-agent" Oct 10 15:08:07 crc kubenswrapper[4788]: E1010 15:08:07.975054 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="proxy-httpd" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975061 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="proxy-httpd" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975427 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-log" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975457 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" containerName="nova-api-api" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975478 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="proxy-httpd" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975501 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="ceilometer-notification-agent" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975517 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="ceilometer-central-agent" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.975529 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" containerName="sg-core" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.977415 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.980350 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.988103 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.988702 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 15:08:07 crc kubenswrapper[4788]: I1010 15:08:07.998680 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.004732 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.004838 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856f37d-ac36-4748-974c-f4ad13b594a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.015320 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2sjbr"] Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.019202 4788 scope.go:117] "RemoveContainer" containerID="75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c" Oct 10 15:08:08 crc kubenswrapper[4788]: E1010 15:08:08.021124 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c\": container with ID starting with 75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c not found: ID does not exist" containerID="75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.021180 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c"} err="failed to get container status \"75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c\": rpc error: code = NotFound desc = could not find container \"75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c\": container with ID starting with 75440a5d6c060b259b05da40a003b0e09c536844eb30bb2db6e6a3af4ca66a2c not found: ID does not exist" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.021210 4788 scope.go:117] "RemoveContainer" containerID="7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e" Oct 10 15:08:08 crc kubenswrapper[4788]: E1010 15:08:08.022373 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e\": container with ID starting with 7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e not found: ID does not exist" containerID="7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.022457 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e"} err="failed to get container status \"7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e\": rpc error: code = NotFound desc = could not find container \"7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e\": container with ID starting with 7a5462246190b4d3b6603ec5f633a8fa93a0a1700ccec9cca7b73c28336ce51e not found: ID does not exist" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.022513 4788 scope.go:117] "RemoveContainer" containerID="89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239" Oct 10 15:08:08 crc kubenswrapper[4788]: E1010 15:08:08.023483 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239\": container with ID starting with 89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239 not found: ID does not exist" containerID="89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.023825 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239"} err="failed to get container status \"89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239\": rpc error: code = NotFound desc = could not find container \"89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239\": container with ID starting with 89cceaf4f8264a9562f505289ddc38c41f776c219b3d0957ee587fcafaac2239 not found: ID does not exist" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.023856 4788 scope.go:117] "RemoveContainer" containerID="d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25" Oct 10 15:08:08 crc kubenswrapper[4788]: E1010 15:08:08.024237 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25\": container with ID starting with d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25 not found: ID does not exist" containerID="d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.024270 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25"} err="failed to get container status \"d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25\": rpc error: code = NotFound desc = could not find container \"d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25\": container with ID starting with d6917ef68ec0062d3da616e8bd2f7a6dced7e1d2072e85be6454120ebbb23d25 not found: ID does not exist" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.024306 4788 scope.go:117] "RemoveContainer" containerID="825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.106074 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z88p8\" (UniqueName: \"kubernetes.io/projected/9bce999d-c0ad-485f-986b-6f6c25004f36-kube-api-access-z88p8\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.106120 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bce999d-c0ad-485f-986b-6f6c25004f36-logs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.106166 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.106223 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-config-data\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.106291 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.106324 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-public-tls-certs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.165961 4788 scope.go:117] "RemoveContainer" containerID="2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.203080 4788 scope.go:117] "RemoveContainer" containerID="825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e" Oct 10 15:08:08 crc kubenswrapper[4788]: E1010 15:08:08.203704 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e\": container with ID starting with 825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e not found: ID does not exist" containerID="825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.203850 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e"} err="failed to get container status \"825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e\": rpc error: code = NotFound desc = could not find container \"825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e\": container with ID starting with 825bb4d689f7ad851683aa1612a7a7871b76312adae900b4c505dba95ff1e74e not found: ID does not exist" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.204010 4788 scope.go:117] "RemoveContainer" containerID="2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7" Oct 10 15:08:08 crc kubenswrapper[4788]: E1010 15:08:08.205395 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7\": container with ID starting with 2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7 not found: ID does not exist" containerID="2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.205515 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7"} err="failed to get container status \"2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7\": rpc error: code = NotFound desc = could not find container \"2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7\": container with ID starting with 2b095c8728e326129c0dd66139449deb3e96e44a22de200b6e3f3c22ca9580f7 not found: ID does not exist" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.208216 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z88p8\" (UniqueName: \"kubernetes.io/projected/9bce999d-c0ad-485f-986b-6f6c25004f36-kube-api-access-z88p8\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.208341 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bce999d-c0ad-485f-986b-6f6c25004f36-logs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.208446 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.208939 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-config-data\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.209107 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.209271 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-public-tls-certs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.210837 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bce999d-c0ad-485f-986b-6f6c25004f36-logs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.212608 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-public-tls-certs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.221759 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.229513 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-config-data\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.232542 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z88p8\" (UniqueName: \"kubernetes.io/projected/9bce999d-c0ad-485f-986b-6f6c25004f36-kube-api-access-z88p8\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.248891 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.256822 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57f5529-f3b7-4620-b81a-4ab31c3a5dda" path="/var/lib/kubelet/pods/a57f5529-f3b7-4620-b81a-4ab31c3a5dda/volumes" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.257788 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.257890 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.265253 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.269111 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.277319 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.279126 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.279544 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.280943 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.412769 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc9dc\" (UniqueName: \"kubernetes.io/projected/4aeeba9b-9894-4293-a673-7e88c6bc80f5-kube-api-access-jc9dc\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.413127 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-run-httpd\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.413272 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.413439 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-log-httpd\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.413628 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.413741 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.413856 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-config-data\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.413958 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-scripts\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.484497 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.516388 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-run-httpd\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.516435 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.516491 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-log-httpd\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.516559 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.516599 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.516624 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-config-data\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.516646 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-scripts\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.516692 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc9dc\" (UniqueName: \"kubernetes.io/projected/4aeeba9b-9894-4293-a673-7e88c6bc80f5-kube-api-access-jc9dc\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.517562 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-log-httpd\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.517801 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-run-httpd\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.523570 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-scripts\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.523632 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.524894 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-config-data\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.528188 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.530552 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.545911 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc9dc\" (UniqueName: \"kubernetes.io/projected/4aeeba9b-9894-4293-a673-7e88c6bc80f5-kube-api-access-jc9dc\") pod \"ceilometer-0\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.665738 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.883606 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2sjbr" event={"ID":"b0efd45c-b649-47ad-a008-349c8380a27a","Type":"ContainerStarted","Data":"1dec38763ab92d684d95ae78a000552b0eab6a7c2f48c775ab127484d305983d"} Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.883660 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2sjbr" event={"ID":"b0efd45c-b649-47ad-a008-349c8380a27a","Type":"ContainerStarted","Data":"5f7d0a1283117cc29668ddf2e390e814af2a9784c3902a29e367593a70af37d6"} Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.910248 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2sjbr" podStartSLOduration=1.910230039 podStartE2EDuration="1.910230039s" podCreationTimestamp="2025-10-10 15:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:08:08.901599935 +0000 UTC m=+1391.351315483" watchObservedRunningTime="2025-10-10 15:08:08.910230039 +0000 UTC m=+1391.359945587" Oct 10 15:08:08 crc kubenswrapper[4788]: I1010 15:08:08.995397 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:09 crc kubenswrapper[4788]: I1010 15:08:09.159575 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:08:09 crc kubenswrapper[4788]: W1010 15:08:09.171520 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aeeba9b_9894_4293_a673_7e88c6bc80f5.slice/crio-9512de2e304f913ec19e124ad1197def2e433a95716f5eb2209fe64db2f17616 WatchSource:0}: Error finding container 9512de2e304f913ec19e124ad1197def2e433a95716f5eb2209fe64db2f17616: Status 404 returned error can't find the container with id 9512de2e304f913ec19e124ad1197def2e433a95716f5eb2209fe64db2f17616 Oct 10 15:08:09 crc kubenswrapper[4788]: I1010 15:08:09.176364 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 15:08:09 crc kubenswrapper[4788]: I1010 15:08:09.897634 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9bce999d-c0ad-485f-986b-6f6c25004f36","Type":"ContainerStarted","Data":"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2"} Oct 10 15:08:09 crc kubenswrapper[4788]: I1010 15:08:09.897984 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9bce999d-c0ad-485f-986b-6f6c25004f36","Type":"ContainerStarted","Data":"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c"} Oct 10 15:08:09 crc kubenswrapper[4788]: I1010 15:08:09.897995 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9bce999d-c0ad-485f-986b-6f6c25004f36","Type":"ContainerStarted","Data":"9f5b608ac8803887efd62717a0744d184e3282128b2febc7309ece682ca32e56"} Oct 10 15:08:09 crc kubenswrapper[4788]: I1010 15:08:09.902105 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerStarted","Data":"b2c6bf83b23d7eb9e36e2d8a2de6b0fcfeb53ed06aa0aaeb6182824c8e84405f"} Oct 10 15:08:09 crc kubenswrapper[4788]: I1010 15:08:09.902161 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerStarted","Data":"9512de2e304f913ec19e124ad1197def2e433a95716f5eb2209fe64db2f17616"} Oct 10 15:08:09 crc kubenswrapper[4788]: I1010 15:08:09.926084 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9260597710000003 podStartE2EDuration="2.926059771s" podCreationTimestamp="2025-10-10 15:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:08:09.919120194 +0000 UTC m=+1392.368835752" watchObservedRunningTime="2025-10-10 15:08:09.926059771 +0000 UTC m=+1392.375775319" Oct 10 15:08:10 crc kubenswrapper[4788]: I1010 15:08:10.250814 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d856f37d-ac36-4748-974c-f4ad13b594a5" path="/var/lib/kubelet/pods/d856f37d-ac36-4748-974c-f4ad13b594a5/volumes" Oct 10 15:08:10 crc kubenswrapper[4788]: I1010 15:08:10.915689 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerStarted","Data":"6065428399b804c4b4b31556aa5431d788e52a568e70173b39bb05a529f66b04"} Oct 10 15:08:11 crc kubenswrapper[4788]: I1010 15:08:11.567472 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:08:11 crc kubenswrapper[4788]: I1010 15:08:11.647964 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hhd8c"] Oct 10 15:08:11 crc kubenswrapper[4788]: I1010 15:08:11.653834 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" podUID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" containerName="dnsmasq-dns" containerID="cri-o://6d0cb49f9707675944733cc27f4c654b8f06a53b69f92ad3276a8956de8ab293" gracePeriod=10 Oct 10 15:08:11 crc kubenswrapper[4788]: I1010 15:08:11.934374 4788 generic.go:334] "Generic (PLEG): container finished" podID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" containerID="6d0cb49f9707675944733cc27f4c654b8f06a53b69f92ad3276a8956de8ab293" exitCode=0 Oct 10 15:08:11 crc kubenswrapper[4788]: I1010 15:08:11.934473 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" event={"ID":"c40f46e4-4197-4b2f-9cf7-c3b3d013838d","Type":"ContainerDied","Data":"6d0cb49f9707675944733cc27f4c654b8f06a53b69f92ad3276a8956de8ab293"} Oct 10 15:08:11 crc kubenswrapper[4788]: I1010 15:08:11.941141 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerStarted","Data":"0a1d2c1f952efc125bdddea9f7888d1cfbf3c24aa07f6a81eb81823168495ca6"} Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.297297 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.480229 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-swift-storage-0\") pod \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.480333 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-sb\") pod \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.480355 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-nb\") pod \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.480424 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frw2b\" (UniqueName: \"kubernetes.io/projected/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-kube-api-access-frw2b\") pod \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.480471 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-config\") pod \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.480533 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-svc\") pod \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\" (UID: \"c40f46e4-4197-4b2f-9cf7-c3b3d013838d\") " Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.485385 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-kube-api-access-frw2b" (OuterVolumeSpecName: "kube-api-access-frw2b") pod "c40f46e4-4197-4b2f-9cf7-c3b3d013838d" (UID: "c40f46e4-4197-4b2f-9cf7-c3b3d013838d"). InnerVolumeSpecName "kube-api-access-frw2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.539258 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c40f46e4-4197-4b2f-9cf7-c3b3d013838d" (UID: "c40f46e4-4197-4b2f-9cf7-c3b3d013838d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.541156 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c40f46e4-4197-4b2f-9cf7-c3b3d013838d" (UID: "c40f46e4-4197-4b2f-9cf7-c3b3d013838d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.542642 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-config" (OuterVolumeSpecName: "config") pod "c40f46e4-4197-4b2f-9cf7-c3b3d013838d" (UID: "c40f46e4-4197-4b2f-9cf7-c3b3d013838d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.545805 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c40f46e4-4197-4b2f-9cf7-c3b3d013838d" (UID: "c40f46e4-4197-4b2f-9cf7-c3b3d013838d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.573086 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c40f46e4-4197-4b2f-9cf7-c3b3d013838d" (UID: "c40f46e4-4197-4b2f-9cf7-c3b3d013838d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.582495 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.582532 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.582542 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.582552 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frw2b\" (UniqueName: \"kubernetes.io/projected/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-kube-api-access-frw2b\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.582565 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.582577 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c40f46e4-4197-4b2f-9cf7-c3b3d013838d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.955829 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" event={"ID":"c40f46e4-4197-4b2f-9cf7-c3b3d013838d","Type":"ContainerDied","Data":"9867e85b57b7659573a6b044235801cdfedbecd6e5a2d59197e8d5b562255a6c"} Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.955898 4788 scope.go:117] "RemoveContainer" containerID="6d0cb49f9707675944733cc27f4c654b8f06a53b69f92ad3276a8956de8ab293" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.956056 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-hhd8c" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.961041 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerStarted","Data":"cc438a9247888adf5836596dc8ce1d20b78dd2a805885fd538540aba1444f53e"} Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.961461 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 15:08:12 crc kubenswrapper[4788]: I1010 15:08:12.998696 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.724862201 podStartE2EDuration="4.99868064s" podCreationTimestamp="2025-10-10 15:08:08 +0000 UTC" firstStartedPulling="2025-10-10 15:08:09.175854306 +0000 UTC m=+1391.625569844" lastFinishedPulling="2025-10-10 15:08:12.449672735 +0000 UTC m=+1394.899388283" observedRunningTime="2025-10-10 15:08:12.992965615 +0000 UTC m=+1395.442681163" watchObservedRunningTime="2025-10-10 15:08:12.99868064 +0000 UTC m=+1395.448396188" Oct 10 15:08:13 crc kubenswrapper[4788]: I1010 15:08:13.008897 4788 scope.go:117] "RemoveContainer" containerID="a27b71c2742a9f61882af118971ebd23910a6cecfff6ae21668a4ef35df711ed" Oct 10 15:08:13 crc kubenswrapper[4788]: I1010 15:08:13.019214 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hhd8c"] Oct 10 15:08:13 crc kubenswrapper[4788]: I1010 15:08:13.025719 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-hhd8c"] Oct 10 15:08:13 crc kubenswrapper[4788]: I1010 15:08:13.973132 4788 generic.go:334] "Generic (PLEG): container finished" podID="b0efd45c-b649-47ad-a008-349c8380a27a" containerID="1dec38763ab92d684d95ae78a000552b0eab6a7c2f48c775ab127484d305983d" exitCode=0 Oct 10 15:08:13 crc kubenswrapper[4788]: I1010 15:08:13.973188 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2sjbr" event={"ID":"b0efd45c-b649-47ad-a008-349c8380a27a","Type":"ContainerDied","Data":"1dec38763ab92d684d95ae78a000552b0eab6a7c2f48c775ab127484d305983d"} Oct 10 15:08:14 crc kubenswrapper[4788]: I1010 15:08:14.246565 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" path="/var/lib/kubelet/pods/c40f46e4-4197-4b2f-9cf7-c3b3d013838d/volumes" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.323167 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.439161 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-scripts\") pod \"b0efd45c-b649-47ad-a008-349c8380a27a\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.439308 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc5wc\" (UniqueName: \"kubernetes.io/projected/b0efd45c-b649-47ad-a008-349c8380a27a-kube-api-access-zc5wc\") pod \"b0efd45c-b649-47ad-a008-349c8380a27a\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.439346 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-combined-ca-bundle\") pod \"b0efd45c-b649-47ad-a008-349c8380a27a\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.439396 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-config-data\") pod \"b0efd45c-b649-47ad-a008-349c8380a27a\" (UID: \"b0efd45c-b649-47ad-a008-349c8380a27a\") " Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.445686 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0efd45c-b649-47ad-a008-349c8380a27a-kube-api-access-zc5wc" (OuterVolumeSpecName: "kube-api-access-zc5wc") pod "b0efd45c-b649-47ad-a008-349c8380a27a" (UID: "b0efd45c-b649-47ad-a008-349c8380a27a"). InnerVolumeSpecName "kube-api-access-zc5wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.446793 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-scripts" (OuterVolumeSpecName: "scripts") pod "b0efd45c-b649-47ad-a008-349c8380a27a" (UID: "b0efd45c-b649-47ad-a008-349c8380a27a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.477060 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0efd45c-b649-47ad-a008-349c8380a27a" (UID: "b0efd45c-b649-47ad-a008-349c8380a27a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.479538 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-config-data" (OuterVolumeSpecName: "config-data") pod "b0efd45c-b649-47ad-a008-349c8380a27a" (UID: "b0efd45c-b649-47ad-a008-349c8380a27a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.541586 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.541631 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc5wc\" (UniqueName: \"kubernetes.io/projected/b0efd45c-b649-47ad-a008-349c8380a27a-kube-api-access-zc5wc\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.541643 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.541652 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0efd45c-b649-47ad-a008-349c8380a27a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.994973 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2sjbr" event={"ID":"b0efd45c-b649-47ad-a008-349c8380a27a","Type":"ContainerDied","Data":"5f7d0a1283117cc29668ddf2e390e814af2a9784c3902a29e367593a70af37d6"} Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.995017 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f7d0a1283117cc29668ddf2e390e814af2a9784c3902a29e367593a70af37d6" Oct 10 15:08:15 crc kubenswrapper[4788]: I1010 15:08:15.995631 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2sjbr" Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.213218 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.213494 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerName="nova-api-log" containerID="cri-o://9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c" gracePeriod=30 Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.213601 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerName="nova-api-api" containerID="cri-o://7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2" gracePeriod=30 Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.256001 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.256558 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2da0778e-f2cc-4897-9996-a6973c974954" containerName="nova-scheduler-scheduler" containerID="cri-o://50b6d48d5eeced1e1b4d152b6f21ebd20a78a742d49a57f9611f4997f62ab50e" gracePeriod=30 Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.286664 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.287136 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-log" containerID="cri-o://9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999" gracePeriod=30 Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.287680 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-metadata" containerID="cri-o://65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c" gracePeriod=30 Oct 10 15:08:16 crc kubenswrapper[4788]: I1010 15:08:16.998193 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.008108 4788 generic.go:334] "Generic (PLEG): container finished" podID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerID="7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2" exitCode=0 Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.008157 4788 generic.go:334] "Generic (PLEG): container finished" podID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerID="9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c" exitCode=143 Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.008213 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9bce999d-c0ad-485f-986b-6f6c25004f36","Type":"ContainerDied","Data":"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2"} Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.008244 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9bce999d-c0ad-485f-986b-6f6c25004f36","Type":"ContainerDied","Data":"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c"} Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.008254 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9bce999d-c0ad-485f-986b-6f6c25004f36","Type":"ContainerDied","Data":"9f5b608ac8803887efd62717a0744d184e3282128b2febc7309ece682ca32e56"} Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.008287 4788 scope.go:117] "RemoveContainer" containerID="7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.008407 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.011461 4788 generic.go:334] "Generic (PLEG): container finished" podID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerID="9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999" exitCode=143 Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.011477 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"de2fb68e-d123-4d2e-8bc2-abae919fb22e","Type":"ContainerDied","Data":"9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999"} Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.046078 4788 scope.go:117] "RemoveContainer" containerID="9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.069518 4788 scope.go:117] "RemoveContainer" containerID="7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2" Oct 10 15:08:17 crc kubenswrapper[4788]: E1010 15:08:17.070123 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2\": container with ID starting with 7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2 not found: ID does not exist" containerID="7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.070203 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2"} err="failed to get container status \"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2\": rpc error: code = NotFound desc = could not find container \"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2\": container with ID starting with 7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2 not found: ID does not exist" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.070236 4788 scope.go:117] "RemoveContainer" containerID="9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c" Oct 10 15:08:17 crc kubenswrapper[4788]: E1010 15:08:17.070756 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c\": container with ID starting with 9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c not found: ID does not exist" containerID="9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.070801 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c"} err="failed to get container status \"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c\": rpc error: code = NotFound desc = could not find container \"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c\": container with ID starting with 9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c not found: ID does not exist" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.070823 4788 scope.go:117] "RemoveContainer" containerID="7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.071103 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2"} err="failed to get container status \"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2\": rpc error: code = NotFound desc = could not find container \"7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2\": container with ID starting with 7dcc8e094addd3ac3e8820442b755654d7491e5e32841d89d51335298efd9db2 not found: ID does not exist" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.071132 4788 scope.go:117] "RemoveContainer" containerID="9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.072314 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c"} err="failed to get container status \"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c\": rpc error: code = NotFound desc = could not find container \"9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c\": container with ID starting with 9392528ef38cb9970a5521901a7e008a6c401d31f2edc09102e24e20251ddf4c not found: ID does not exist" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.085616 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-internal-tls-certs\") pod \"9bce999d-c0ad-485f-986b-6f6c25004f36\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.085773 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z88p8\" (UniqueName: \"kubernetes.io/projected/9bce999d-c0ad-485f-986b-6f6c25004f36-kube-api-access-z88p8\") pod \"9bce999d-c0ad-485f-986b-6f6c25004f36\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.085867 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bce999d-c0ad-485f-986b-6f6c25004f36-logs\") pod \"9bce999d-c0ad-485f-986b-6f6c25004f36\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.085936 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-public-tls-certs\") pod \"9bce999d-c0ad-485f-986b-6f6c25004f36\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.086080 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-config-data\") pod \"9bce999d-c0ad-485f-986b-6f6c25004f36\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.086137 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-combined-ca-bundle\") pod \"9bce999d-c0ad-485f-986b-6f6c25004f36\" (UID: \"9bce999d-c0ad-485f-986b-6f6c25004f36\") " Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.086796 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bce999d-c0ad-485f-986b-6f6c25004f36-logs" (OuterVolumeSpecName: "logs") pod "9bce999d-c0ad-485f-986b-6f6c25004f36" (UID: "9bce999d-c0ad-485f-986b-6f6c25004f36"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.092099 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bce999d-c0ad-485f-986b-6f6c25004f36-kube-api-access-z88p8" (OuterVolumeSpecName: "kube-api-access-z88p8") pod "9bce999d-c0ad-485f-986b-6f6c25004f36" (UID: "9bce999d-c0ad-485f-986b-6f6c25004f36"). InnerVolumeSpecName "kube-api-access-z88p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.116531 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9bce999d-c0ad-485f-986b-6f6c25004f36" (UID: "9bce999d-c0ad-485f-986b-6f6c25004f36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.117688 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-config-data" (OuterVolumeSpecName: "config-data") pod "9bce999d-c0ad-485f-986b-6f6c25004f36" (UID: "9bce999d-c0ad-485f-986b-6f6c25004f36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.144002 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9bce999d-c0ad-485f-986b-6f6c25004f36" (UID: "9bce999d-c0ad-485f-986b-6f6c25004f36"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.144940 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9bce999d-c0ad-485f-986b-6f6c25004f36" (UID: "9bce999d-c0ad-485f-986b-6f6c25004f36"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.188561 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z88p8\" (UniqueName: \"kubernetes.io/projected/9bce999d-c0ad-485f-986b-6f6c25004f36-kube-api-access-z88p8\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.188610 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bce999d-c0ad-485f-986b-6f6c25004f36-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.188628 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.188639 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.188650 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.188660 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bce999d-c0ad-485f-986b-6f6c25004f36-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.409322 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.429879 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.440109 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:17 crc kubenswrapper[4788]: E1010 15:08:17.440727 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" containerName="dnsmasq-dns" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.440752 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" containerName="dnsmasq-dns" Oct 10 15:08:17 crc kubenswrapper[4788]: E1010 15:08:17.440765 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerName="nova-api-log" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.440773 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerName="nova-api-log" Oct 10 15:08:17 crc kubenswrapper[4788]: E1010 15:08:17.440802 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerName="nova-api-api" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.440809 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerName="nova-api-api" Oct 10 15:08:17 crc kubenswrapper[4788]: E1010 15:08:17.440834 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" containerName="init" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.440842 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" containerName="init" Oct 10 15:08:17 crc kubenswrapper[4788]: E1010 15:08:17.440863 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0efd45c-b649-47ad-a008-349c8380a27a" containerName="nova-manage" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.440870 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0efd45c-b649-47ad-a008-349c8380a27a" containerName="nova-manage" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.441197 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0efd45c-b649-47ad-a008-349c8380a27a" containerName="nova-manage" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.441218 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerName="nova-api-api" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.441237 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c40f46e4-4197-4b2f-9cf7-c3b3d013838d" containerName="dnsmasq-dns" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.441253 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" containerName="nova-api-log" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.442757 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.445177 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.445506 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.445715 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.465996 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.596982 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-config-data\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.597051 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.597243 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcxnq\" (UniqueName: \"kubernetes.io/projected/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-kube-api-access-bcxnq\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.597294 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.597321 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.597353 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-logs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.698784 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcxnq\" (UniqueName: \"kubernetes.io/projected/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-kube-api-access-bcxnq\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.698882 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.698911 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.698943 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-logs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.699003 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-config-data\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.699030 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.700435 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-logs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.703640 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.703955 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-config-data\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.703969 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.704945 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-public-tls-certs\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.722588 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcxnq\" (UniqueName: \"kubernetes.io/projected/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-kube-api-access-bcxnq\") pod \"nova-api-0\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " pod="openstack/nova-api-0" Oct 10 15:08:17 crc kubenswrapper[4788]: I1010 15:08:17.778382 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.027340 4788 generic.go:334] "Generic (PLEG): container finished" podID="2da0778e-f2cc-4897-9996-a6973c974954" containerID="50b6d48d5eeced1e1b4d152b6f21ebd20a78a742d49a57f9611f4997f62ab50e" exitCode=0 Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.027777 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2da0778e-f2cc-4897-9996-a6973c974954","Type":"ContainerDied","Data":"50b6d48d5eeced1e1b4d152b6f21ebd20a78a742d49a57f9611f4997f62ab50e"} Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.226622 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:08:18 crc kubenswrapper[4788]: W1010 15:08:18.229971 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b7f2c05_93d2_4139_95ea_c36bffcecdb8.slice/crio-5a890daae46b4f28ec56c1ab8637d5a49495dfd1c24de91adea89c800024351a WatchSource:0}: Error finding container 5a890daae46b4f28ec56c1ab8637d5a49495dfd1c24de91adea89c800024351a: Status 404 returned error can't find the container with id 5a890daae46b4f28ec56c1ab8637d5a49495dfd1c24de91adea89c800024351a Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.257126 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bce999d-c0ad-485f-986b-6f6c25004f36" path="/var/lib/kubelet/pods/9bce999d-c0ad-485f-986b-6f6c25004f36/volumes" Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.282830 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.414803 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n766m\" (UniqueName: \"kubernetes.io/projected/2da0778e-f2cc-4897-9996-a6973c974954-kube-api-access-n766m\") pod \"2da0778e-f2cc-4897-9996-a6973c974954\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.414947 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-combined-ca-bundle\") pod \"2da0778e-f2cc-4897-9996-a6973c974954\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.415101 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-config-data\") pod \"2da0778e-f2cc-4897-9996-a6973c974954\" (UID: \"2da0778e-f2cc-4897-9996-a6973c974954\") " Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.423531 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da0778e-f2cc-4897-9996-a6973c974954-kube-api-access-n766m" (OuterVolumeSpecName: "kube-api-access-n766m") pod "2da0778e-f2cc-4897-9996-a6973c974954" (UID: "2da0778e-f2cc-4897-9996-a6973c974954"). InnerVolumeSpecName "kube-api-access-n766m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.454809 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-config-data" (OuterVolumeSpecName: "config-data") pod "2da0778e-f2cc-4897-9996-a6973c974954" (UID: "2da0778e-f2cc-4897-9996-a6973c974954"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.454896 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2da0778e-f2cc-4897-9996-a6973c974954" (UID: "2da0778e-f2cc-4897-9996-a6973c974954"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.517940 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.517969 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da0778e-f2cc-4897-9996-a6973c974954-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:18 crc kubenswrapper[4788]: I1010 15:08:18.517981 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n766m\" (UniqueName: \"kubernetes.io/projected/2da0778e-f2cc-4897-9996-a6973c974954-kube-api-access-n766m\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.039251 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b7f2c05-93d2-4139-95ea-c36bffcecdb8","Type":"ContainerStarted","Data":"598ec17f8706ac8a7613481f8dfaf31904983cb92d6010eefbf1ac47efcbb014"} Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.040371 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b7f2c05-93d2-4139-95ea-c36bffcecdb8","Type":"ContainerStarted","Data":"79447d873480f8db4080543745a5c249b2cbec57b2c886298cf38288010480c5"} Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.040492 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b7f2c05-93d2-4139-95ea-c36bffcecdb8","Type":"ContainerStarted","Data":"5a890daae46b4f28ec56c1ab8637d5a49495dfd1c24de91adea89c800024351a"} Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.041059 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2da0778e-f2cc-4897-9996-a6973c974954","Type":"ContainerDied","Data":"31dfedc38abb192f7849761f63090ac4313d210060076fdee2eb412af7d5f48a"} Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.041112 4788 scope.go:117] "RemoveContainer" containerID="50b6d48d5eeced1e1b4d152b6f21ebd20a78a742d49a57f9611f4997f62ab50e" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.041216 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.059748 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.059726385 podStartE2EDuration="2.059726385s" podCreationTimestamp="2025-10-10 15:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:08:19.055124661 +0000 UTC m=+1401.504840219" watchObservedRunningTime="2025-10-10 15:08:19.059726385 +0000 UTC m=+1401.509441933" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.084176 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.092374 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.108676 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:08:19 crc kubenswrapper[4788]: E1010 15:08:19.109350 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da0778e-f2cc-4897-9996-a6973c974954" containerName="nova-scheduler-scheduler" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.109381 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da0778e-f2cc-4897-9996-a6973c974954" containerName="nova-scheduler-scheduler" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.109747 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da0778e-f2cc-4897-9996-a6973c974954" containerName="nova-scheduler-scheduler" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.110915 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.117555 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.122382 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.235438 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-config-data\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.235759 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88574\" (UniqueName: \"kubernetes.io/projected/4fad8d4a-3435-4647-b10d-51951c64e49a-kube-api-access-88574\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.235871 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.338366 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.338884 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-config-data\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.339326 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88574\" (UniqueName: \"kubernetes.io/projected/4fad8d4a-3435-4647-b10d-51951c64e49a-kube-api-access-88574\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.342397 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-config-data\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.361096 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.364826 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88574\" (UniqueName: \"kubernetes.io/projected/4fad8d4a-3435-4647-b10d-51951c64e49a-kube-api-access-88574\") pod \"nova-scheduler-0\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.431124 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.443347 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:51122->10.217.0.192:8775: read: connection reset by peer" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.443448 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:51138->10.217.0.192:8775: read: connection reset by peer" Oct 10 15:08:19 crc kubenswrapper[4788]: W1010 15:08:19.933838 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fad8d4a_3435_4647_b10d_51951c64e49a.slice/crio-48a796b250ee9e5c80cca00d0a66e4bafe6672c6c42fe572bb77dfc8eadfcfa4 WatchSource:0}: Error finding container 48a796b250ee9e5c80cca00d0a66e4bafe6672c6c42fe572bb77dfc8eadfcfa4: Status 404 returned error can't find the container with id 48a796b250ee9e5c80cca00d0a66e4bafe6672c6c42fe572bb77dfc8eadfcfa4 Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.941280 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:08:19 crc kubenswrapper[4788]: I1010 15:08:19.946298 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.055902 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4fad8d4a-3435-4647-b10d-51951c64e49a","Type":"ContainerStarted","Data":"48a796b250ee9e5c80cca00d0a66e4bafe6672c6c42fe572bb77dfc8eadfcfa4"} Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.059072 4788 generic.go:334] "Generic (PLEG): container finished" podID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerID="65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c" exitCode=0 Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.059133 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"de2fb68e-d123-4d2e-8bc2-abae919fb22e","Type":"ContainerDied","Data":"65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c"} Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.059179 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"de2fb68e-d123-4d2e-8bc2-abae919fb22e","Type":"ContainerDied","Data":"f87cb49e093f9773649f15564dabd4a93c1852334703cef80c55c84c20857729"} Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.059195 4788 scope.go:117] "RemoveContainer" containerID="65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.059312 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.059970 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2fb68e-d123-4d2e-8bc2-abae919fb22e-logs\") pod \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.060125 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftg6r\" (UniqueName: \"kubernetes.io/projected/de2fb68e-d123-4d2e-8bc2-abae919fb22e-kube-api-access-ftg6r\") pod \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.060199 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-nova-metadata-tls-certs\") pod \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.060240 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-config-data\") pod \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.060314 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-combined-ca-bundle\") pod \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\" (UID: \"de2fb68e-d123-4d2e-8bc2-abae919fb22e\") " Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.060660 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2fb68e-d123-4d2e-8bc2-abae919fb22e-logs" (OuterVolumeSpecName: "logs") pod "de2fb68e-d123-4d2e-8bc2-abae919fb22e" (UID: "de2fb68e-d123-4d2e-8bc2-abae919fb22e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.060865 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de2fb68e-d123-4d2e-8bc2-abae919fb22e-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.074270 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2fb68e-d123-4d2e-8bc2-abae919fb22e-kube-api-access-ftg6r" (OuterVolumeSpecName: "kube-api-access-ftg6r") pod "de2fb68e-d123-4d2e-8bc2-abae919fb22e" (UID: "de2fb68e-d123-4d2e-8bc2-abae919fb22e"). InnerVolumeSpecName "kube-api-access-ftg6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.091463 4788 scope.go:117] "RemoveContainer" containerID="9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.101743 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-config-data" (OuterVolumeSpecName: "config-data") pod "de2fb68e-d123-4d2e-8bc2-abae919fb22e" (UID: "de2fb68e-d123-4d2e-8bc2-abae919fb22e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.107608 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de2fb68e-d123-4d2e-8bc2-abae919fb22e" (UID: "de2fb68e-d123-4d2e-8bc2-abae919fb22e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.114753 4788 scope.go:117] "RemoveContainer" containerID="65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c" Oct 10 15:08:20 crc kubenswrapper[4788]: E1010 15:08:20.115245 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c\": container with ID starting with 65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c not found: ID does not exist" containerID="65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.115285 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c"} err="failed to get container status \"65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c\": rpc error: code = NotFound desc = could not find container \"65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c\": container with ID starting with 65d4452d619b6dc1f32a1620ac863662c29b9ec32af546d5701e118d8d88c60c not found: ID does not exist" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.115312 4788 scope.go:117] "RemoveContainer" containerID="9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999" Oct 10 15:08:20 crc kubenswrapper[4788]: E1010 15:08:20.115546 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999\": container with ID starting with 9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999 not found: ID does not exist" containerID="9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.115568 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999"} err="failed to get container status \"9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999\": rpc error: code = NotFound desc = could not find container \"9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999\": container with ID starting with 9ade4f3707b6dc65fd6c27ba5de0f6fc963be135e8124dadde4e703bff64f999 not found: ID does not exist" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.140043 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "de2fb68e-d123-4d2e-8bc2-abae919fb22e" (UID: "de2fb68e-d123-4d2e-8bc2-abae919fb22e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.162557 4788 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.162592 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.162604 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2fb68e-d123-4d2e-8bc2-abae919fb22e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.162613 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftg6r\" (UniqueName: \"kubernetes.io/projected/de2fb68e-d123-4d2e-8bc2-abae919fb22e-kube-api-access-ftg6r\") on node \"crc\" DevicePath \"\"" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.246797 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2da0778e-f2cc-4897-9996-a6973c974954" path="/var/lib/kubelet/pods/2da0778e-f2cc-4897-9996-a6973c974954/volumes" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.424005 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.437444 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.459472 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:08:20 crc kubenswrapper[4788]: E1010 15:08:20.459981 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-metadata" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.460007 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-metadata" Oct 10 15:08:20 crc kubenswrapper[4788]: E1010 15:08:20.460035 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-log" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.460044 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-log" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.460321 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-log" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.460342 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" containerName="nova-metadata-metadata" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.461692 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.465820 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.466070 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.473973 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.571586 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d2fx\" (UniqueName: \"kubernetes.io/projected/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-kube-api-access-2d2fx\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.571967 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-config-data\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.572008 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.572034 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.572057 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-logs\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.673465 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-config-data\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.673527 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.673550 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.673575 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-logs\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.673676 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d2fx\" (UniqueName: \"kubernetes.io/projected/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-kube-api-access-2d2fx\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.674130 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-logs\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.679935 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-config-data\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.679932 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.682390 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.692335 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d2fx\" (UniqueName: \"kubernetes.io/projected/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-kube-api-access-2d2fx\") pod \"nova-metadata-0\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " pod="openstack/nova-metadata-0" Oct 10 15:08:20 crc kubenswrapper[4788]: I1010 15:08:20.787600 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:08:21 crc kubenswrapper[4788]: I1010 15:08:21.082883 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4fad8d4a-3435-4647-b10d-51951c64e49a","Type":"ContainerStarted","Data":"edf59e020f4a2e0970646b96aab104bff00861373f9c92c494f167dd813d4c13"} Oct 10 15:08:21 crc kubenswrapper[4788]: I1010 15:08:21.105012 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.104995769 podStartE2EDuration="2.104995769s" podCreationTimestamp="2025-10-10 15:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:08:21.099679846 +0000 UTC m=+1403.549395394" watchObservedRunningTime="2025-10-10 15:08:21.104995769 +0000 UTC m=+1403.554711317" Oct 10 15:08:21 crc kubenswrapper[4788]: I1010 15:08:21.264510 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:08:21 crc kubenswrapper[4788]: W1010 15:08:21.269337 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2d4fc78_0ea7_49fe_9da3_caed0438e9f8.slice/crio-1ce4d467f7c9d92a86b58fa6239577d1a24246e94ca1434de7ba399a191e8d73 WatchSource:0}: Error finding container 1ce4d467f7c9d92a86b58fa6239577d1a24246e94ca1434de7ba399a191e8d73: Status 404 returned error can't find the container with id 1ce4d467f7c9d92a86b58fa6239577d1a24246e94ca1434de7ba399a191e8d73 Oct 10 15:08:22 crc kubenswrapper[4788]: I1010 15:08:22.096630 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8","Type":"ContainerStarted","Data":"fab44d16043a2503f608627193df1f5950c0f9017dc038519c69453dcac75e2f"} Oct 10 15:08:22 crc kubenswrapper[4788]: I1010 15:08:22.096926 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8","Type":"ContainerStarted","Data":"5bc7218d76c149b7925383f9d5d6ab8180d2bb5b1b399cf25e15bef72a048be8"} Oct 10 15:08:22 crc kubenswrapper[4788]: I1010 15:08:22.096936 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8","Type":"ContainerStarted","Data":"1ce4d467f7c9d92a86b58fa6239577d1a24246e94ca1434de7ba399a191e8d73"} Oct 10 15:08:22 crc kubenswrapper[4788]: I1010 15:08:22.122777 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.122759115 podStartE2EDuration="2.122759115s" podCreationTimestamp="2025-10-10 15:08:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:08:22.119427335 +0000 UTC m=+1404.569142893" watchObservedRunningTime="2025-10-10 15:08:22.122759115 +0000 UTC m=+1404.572474663" Oct 10 15:08:22 crc kubenswrapper[4788]: I1010 15:08:22.247445 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de2fb68e-d123-4d2e-8bc2-abae919fb22e" path="/var/lib/kubelet/pods/de2fb68e-d123-4d2e-8bc2-abae919fb22e/volumes" Oct 10 15:08:24 crc kubenswrapper[4788]: I1010 15:08:24.431572 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 15:08:25 crc kubenswrapper[4788]: I1010 15:08:25.788577 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 15:08:25 crc kubenswrapper[4788]: I1010 15:08:25.788673 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 15:08:27 crc kubenswrapper[4788]: I1010 15:08:27.779951 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 15:08:27 crc kubenswrapper[4788]: I1010 15:08:27.780497 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 15:08:28 crc kubenswrapper[4788]: I1010 15:08:28.798471 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 15:08:28 crc kubenswrapper[4788]: I1010 15:08:28.798457 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 15:08:29 crc kubenswrapper[4788]: I1010 15:08:29.407392 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:08:29 crc kubenswrapper[4788]: I1010 15:08:29.407478 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:08:29 crc kubenswrapper[4788]: I1010 15:08:29.431379 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 15:08:29 crc kubenswrapper[4788]: I1010 15:08:29.467574 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 15:08:30 crc kubenswrapper[4788]: I1010 15:08:30.274191 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 15:08:30 crc kubenswrapper[4788]: I1010 15:08:30.789293 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 15:08:30 crc kubenswrapper[4788]: I1010 15:08:30.789350 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 15:08:31 crc kubenswrapper[4788]: I1010 15:08:31.805577 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 15:08:31 crc kubenswrapper[4788]: I1010 15:08:31.805614 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 10 15:08:37 crc kubenswrapper[4788]: I1010 15:08:37.792062 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 15:08:37 crc kubenswrapper[4788]: I1010 15:08:37.794536 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 15:08:37 crc kubenswrapper[4788]: I1010 15:08:37.796177 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 15:08:37 crc kubenswrapper[4788]: I1010 15:08:37.796223 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 15:08:37 crc kubenswrapper[4788]: I1010 15:08:37.807505 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 15:08:37 crc kubenswrapper[4788]: I1010 15:08:37.809177 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 15:08:38 crc kubenswrapper[4788]: I1010 15:08:38.680580 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 15:08:40 crc kubenswrapper[4788]: I1010 15:08:40.795202 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 15:08:40 crc kubenswrapper[4788]: I1010 15:08:40.796342 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 15:08:40 crc kubenswrapper[4788]: I1010 15:08:40.808702 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 15:08:40 crc kubenswrapper[4788]: I1010 15:08:40.811307 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 15:08:59 crc kubenswrapper[4788]: I1010 15:08:59.406683 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:08:59 crc kubenswrapper[4788]: I1010 15:08:59.407227 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.162128 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.163045 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="d84e4226-ac3f-4f6f-870b-c795e206854e" containerName="openstackclient" containerID="cri-o://62068765dbbe9a789befff7b25adc8fe43b248f71c27fd2edbc0898452134000" gracePeriod=2 Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.198256 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.335759 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-wpgvn"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.396300 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.396585 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="ovn-northd" containerID="cri-o://ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" gracePeriod=30 Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.396706 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="openstack-network-exporter" containerID="cri-o://335000a0690e32f11a92cfd8dfc8fe055195d80c307ea823dc4da83c8237d39d" gracePeriod=30 Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.407491 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-n84gv"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.437387 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-7m7sz"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.437652 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-7m7sz" podUID="953774da-73b1-4f46-aa87-8deba0bcc077" containerName="openstack-network-exporter" containerID="cri-o://69a85b38a0e4154979d9d710277ca0a1c7a9d6c594026db5c938404d7dc3173b" gracePeriod=30 Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.462360 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican59ec-account-delete-nn85b"] Oct 10 15:09:03 crc kubenswrapper[4788]: E1010 15:09:03.462796 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84e4226-ac3f-4f6f-870b-c795e206854e" containerName="openstackclient" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.462809 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84e4226-ac3f-4f6f-870b-c795e206854e" containerName="openstackclient" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.462997 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84e4226-ac3f-4f6f-870b-c795e206854e" containerName="openstackclient" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.463680 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican59ec-account-delete-nn85b" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.492211 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.527273 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement4ba5-account-delete-ljq8m"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.529004 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement4ba5-account-delete-ljq8m" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.548409 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6kwj\" (UniqueName: \"kubernetes.io/projected/5bfee301-605c-4fb8-992e-8911bfc7206f-kube-api-access-t6kwj\") pod \"barbican59ec-account-delete-nn85b\" (UID: \"5bfee301-605c-4fb8-992e-8911bfc7206f\") " pod="openstack/barbican59ec-account-delete-nn85b" Oct 10 15:09:03 crc kubenswrapper[4788]: E1010 15:09:03.548697 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 10 15:09:03 crc kubenswrapper[4788]: E1010 15:09:03.548747 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data podName:dcf4b178-75ee-47fb-81e4-e6b277eb1790 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:04.048731048 +0000 UTC m=+1446.498446596 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data") pod "rabbitmq-server-0" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790") : configmap "rabbitmq-config-data" not found Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.580675 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican59ec-account-delete-nn85b"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.622314 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement4ba5-account-delete-ljq8m"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.657206 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6kwj\" (UniqueName: \"kubernetes.io/projected/5bfee301-605c-4fb8-992e-8911bfc7206f-kube-api-access-t6kwj\") pod \"barbican59ec-account-delete-nn85b\" (UID: \"5bfee301-605c-4fb8-992e-8911bfc7206f\") " pod="openstack/barbican59ec-account-delete-nn85b" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.743260 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinderb986-account-delete-tvgzp"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.745404 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderb986-account-delete-tvgzp" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.783289 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6kwj\" (UniqueName: \"kubernetes.io/projected/5bfee301-605c-4fb8-992e-8911bfc7206f-kube-api-access-t6kwj\") pod \"barbican59ec-account-delete-nn85b\" (UID: \"5bfee301-605c-4fb8-992e-8911bfc7206f\") " pod="openstack/barbican59ec-account-delete-nn85b" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.782554 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcgx6\" (UniqueName: \"kubernetes.io/projected/7810b9ee-1996-4535-bcac-5ffa8781295b-kube-api-access-fcgx6\") pod \"cinderb986-account-delete-tvgzp\" (UID: \"7810b9ee-1996-4535-bcac-5ffa8781295b\") " pod="openstack/cinderb986-account-delete-tvgzp" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.789352 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjpfm\" (UniqueName: \"kubernetes.io/projected/024ba580-feec-4b2a-a836-942126b022ae-kube-api-access-xjpfm\") pod \"placement4ba5-account-delete-ljq8m\" (UID: \"024ba580-feec-4b2a-a836-942126b022ae\") " pod="openstack/placement4ba5-account-delete-ljq8m" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.800440 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderb986-account-delete-tvgzp"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.834517 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican59ec-account-delete-nn85b" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.843485 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.897038 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcgx6\" (UniqueName: \"kubernetes.io/projected/7810b9ee-1996-4535-bcac-5ffa8781295b-kube-api-access-fcgx6\") pod \"cinderb986-account-delete-tvgzp\" (UID: \"7810b9ee-1996-4535-bcac-5ffa8781295b\") " pod="openstack/cinderb986-account-delete-tvgzp" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.897101 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjpfm\" (UniqueName: \"kubernetes.io/projected/024ba580-feec-4b2a-a836-942126b022ae-kube-api-access-xjpfm\") pod \"placement4ba5-account-delete-ljq8m\" (UID: \"024ba580-feec-4b2a-a836-942126b022ae\") " pod="openstack/placement4ba5-account-delete-ljq8m" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.920075 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance9145-account-delete-kblv8"] Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.929463 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance9145-account-delete-kblv8" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.954297 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcgx6\" (UniqueName: \"kubernetes.io/projected/7810b9ee-1996-4535-bcac-5ffa8781295b-kube-api-access-fcgx6\") pod \"cinderb986-account-delete-tvgzp\" (UID: \"7810b9ee-1996-4535-bcac-5ffa8781295b\") " pod="openstack/cinderb986-account-delete-tvgzp" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.988908 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjpfm\" (UniqueName: \"kubernetes.io/projected/024ba580-feec-4b2a-a836-942126b022ae-kube-api-access-xjpfm\") pod \"placement4ba5-account-delete-ljq8m\" (UID: \"024ba580-feec-4b2a-a836-942126b022ae\") " pod="openstack/placement4ba5-account-delete-ljq8m" Oct 10 15:09:03 crc kubenswrapper[4788]: I1010 15:09:03.989776 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance9145-account-delete-kblv8"] Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.011570 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.012505 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data podName:a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c nodeName:}" failed. No retries permitted until 2025-10-10 15:09:04.512492164 +0000 UTC m=+1446.962207712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data") pod "rabbitmq-cell1-server-0" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c") : configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.039577 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronab26-account-delete-g2dnr"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.041187 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronab26-account-delete-g2dnr" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.067044 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronab26-account-delete-g2dnr"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.097481 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi96b1-account-delete-pm4dv"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.099014 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi96b1-account-delete-pm4dv" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.107797 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.108187 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="openstack-network-exporter" containerID="cri-o://61dd4ebbf802647ebbc1de9086c92e02600921dcbef187e7ed7f85e677087139" gracePeriod=300 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.110873 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderb986-account-delete-tvgzp" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.114575 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg9gc\" (UniqueName: \"kubernetes.io/projected/42d4825c-6e5c-4483-ae09-657903e6acb0-kube-api-access-kg9gc\") pod \"glance9145-account-delete-kblv8\" (UID: \"42d4825c-6e5c-4483-ae09-657903e6acb0\") " pod="openstack/glance9145-account-delete-kblv8" Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.114842 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.114892 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data podName:dcf4b178-75ee-47fb-81e4-e6b277eb1790 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:05.114878353 +0000 UTC m=+1447.564593891 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data") pod "rabbitmq-server-0" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790") : configmap "rabbitmq-config-data" not found Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.170907 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi96b1-account-delete-pm4dv"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.221777 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg9gc\" (UniqueName: \"kubernetes.io/projected/42d4825c-6e5c-4483-ae09-657903e6acb0-kube-api-access-kg9gc\") pod \"glance9145-account-delete-kblv8\" (UID: \"42d4825c-6e5c-4483-ae09-657903e6acb0\") " pod="openstack/glance9145-account-delete-kblv8" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.237171 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vchst\" (UniqueName: \"kubernetes.io/projected/88f8c7ed-d478-4bae-ae23-4bffcb16abf6-kube-api-access-vchst\") pod \"neutronab26-account-delete-g2dnr\" (UID: \"88f8c7ed-d478-4bae-ae23-4bffcb16abf6\") " pod="openstack/neutronab26-account-delete-g2dnr" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.237257 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz8sx\" (UniqueName: \"kubernetes.io/projected/0d375458-0dc4-45a2-abf9-c7376510d2f6-kube-api-access-jz8sx\") pod \"novaapi96b1-account-delete-pm4dv\" (UID: \"0d375458-0dc4-45a2-abf9-c7376510d2f6\") " pod="openstack/novaapi96b1-account-delete-pm4dv" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.252082 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement4ba5-account-delete-ljq8m" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.260259 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg9gc\" (UniqueName: \"kubernetes.io/projected/42d4825c-6e5c-4483-ae09-657903e6acb0-kube-api-access-kg9gc\") pod \"glance9145-account-delete-kblv8\" (UID: \"42d4825c-6e5c-4483-ae09-657903e6acb0\") " pod="openstack/glance9145-account-delete-kblv8" Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.350863 4788 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-n84gv" message="Exiting ovn-controller (1) " Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.350915 4788 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-n84gv" podUID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" containerName="ovn-controller" containerID="cri-o://94627d4d5603047c40017358c990871e5690c6c59e30cc3be499dae8b6bf6332" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.350952 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-n84gv" podUID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" containerName="ovn-controller" containerID="cri-o://94627d4d5603047c40017358c990871e5690c6c59e30cc3be499dae8b6bf6332" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.358480 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vchst\" (UniqueName: \"kubernetes.io/projected/88f8c7ed-d478-4bae-ae23-4bffcb16abf6-kube-api-access-vchst\") pod \"neutronab26-account-delete-g2dnr\" (UID: \"88f8c7ed-d478-4bae-ae23-4bffcb16abf6\") " pod="openstack/neutronab26-account-delete-g2dnr" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.358550 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz8sx\" (UniqueName: \"kubernetes.io/projected/0d375458-0dc4-45a2-abf9-c7376510d2f6-kube-api-access-jz8sx\") pod \"novaapi96b1-account-delete-pm4dv\" (UID: \"0d375458-0dc4-45a2-abf9-c7376510d2f6\") " pod="openstack/novaapi96b1-account-delete-pm4dv" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.461672 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz8sx\" (UniqueName: \"kubernetes.io/projected/0d375458-0dc4-45a2-abf9-c7376510d2f6-kube-api-access-jz8sx\") pod \"novaapi96b1-account-delete-pm4dv\" (UID: \"0d375458-0dc4-45a2-abf9-c7376510d2f6\") " pod="openstack/novaapi96b1-account-delete-pm4dv" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.463488 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vchst\" (UniqueName: \"kubernetes.io/projected/88f8c7ed-d478-4bae-ae23-4bffcb16abf6-kube-api-access-vchst\") pod \"neutronab26-account-delete-g2dnr\" (UID: \"88f8c7ed-d478-4bae-ae23-4bffcb16abf6\") " pod="openstack/neutronab26-account-delete-g2dnr" Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.567183 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.567248 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data podName:a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c nodeName:}" failed. No retries permitted until 2025-10-10 15:09:05.567231183 +0000 UTC m=+1448.016946731 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data") pod "rabbitmq-cell1-server-0" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c") : configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.639124 4788 generic.go:334] "Generic (PLEG): container finished" podID="4b927859-0238-46f5-90aa-4afa094c434e" containerID="61dd4ebbf802647ebbc1de9086c92e02600921dcbef187e7ed7f85e677087139" exitCode=2 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.644542 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7m7sz_953774da-73b1-4f46-aa87-8deba0bcc077/openstack-network-exporter/0.log" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.644605 4788 generic.go:334] "Generic (PLEG): container finished" podID="953774da-73b1-4f46-aa87-8deba0bcc077" containerID="69a85b38a0e4154979d9d710277ca0a1c7a9d6c594026db5c938404d7dc3173b" exitCode=2 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.664304 4788 generic.go:334] "Generic (PLEG): container finished" podID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerID="335000a0690e32f11a92cfd8dfc8fe055195d80c307ea823dc4da83c8237d39d" exitCode=2 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.669971 4788 generic.go:334] "Generic (PLEG): container finished" podID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" containerID="94627d4d5603047c40017358c990871e5690c6c59e30cc3be499dae8b6bf6332" exitCode=0 Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.801841 4788 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 10 15:09:04 crc kubenswrapper[4788]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 10 15:09:04 crc kubenswrapper[4788]: + source /usr/local/bin/container-scripts/functions Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNBridge=br-int Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNRemote=tcp:localhost:6642 Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNEncapType=geneve Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNAvailabilityZones= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ EnableChassisAsGateway=true Oct 10 15:09:04 crc kubenswrapper[4788]: ++ PhysicalNetworks= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNHostName= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 10 15:09:04 crc kubenswrapper[4788]: ++ ovs_dir=/var/lib/openvswitch Oct 10 15:09:04 crc kubenswrapper[4788]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 10 15:09:04 crc kubenswrapper[4788]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 10 15:09:04 crc kubenswrapper[4788]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + sleep 0.5 Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + sleep 0.5 Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + cleanup_ovsdb_server_semaphore Oct 10 15:09:04 crc kubenswrapper[4788]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 10 15:09:04 crc kubenswrapper[4788]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 10 15:09:04 crc kubenswrapper[4788]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-wpgvn" message=< Oct 10 15:09:04 crc kubenswrapper[4788]: Exiting ovsdb-server (5) [ OK ] Oct 10 15:09:04 crc kubenswrapper[4788]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 10 15:09:04 crc kubenswrapper[4788]: + source /usr/local/bin/container-scripts/functions Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNBridge=br-int Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNRemote=tcp:localhost:6642 Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNEncapType=geneve Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNAvailabilityZones= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ EnableChassisAsGateway=true Oct 10 15:09:04 crc kubenswrapper[4788]: ++ PhysicalNetworks= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNHostName= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 10 15:09:04 crc kubenswrapper[4788]: ++ ovs_dir=/var/lib/openvswitch Oct 10 15:09:04 crc kubenswrapper[4788]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 10 15:09:04 crc kubenswrapper[4788]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 10 15:09:04 crc kubenswrapper[4788]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + sleep 0.5 Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + sleep 0.5 Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + cleanup_ovsdb_server_semaphore Oct 10 15:09:04 crc kubenswrapper[4788]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 10 15:09:04 crc kubenswrapper[4788]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 10 15:09:04 crc kubenswrapper[4788]: > Oct 10 15:09:04 crc kubenswrapper[4788]: E1010 15:09:04.801879 4788 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 10 15:09:04 crc kubenswrapper[4788]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 10 15:09:04 crc kubenswrapper[4788]: + source /usr/local/bin/container-scripts/functions Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNBridge=br-int Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNRemote=tcp:localhost:6642 Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNEncapType=geneve Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNAvailabilityZones= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ EnableChassisAsGateway=true Oct 10 15:09:04 crc kubenswrapper[4788]: ++ PhysicalNetworks= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ OVNHostName= Oct 10 15:09:04 crc kubenswrapper[4788]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 10 15:09:04 crc kubenswrapper[4788]: ++ ovs_dir=/var/lib/openvswitch Oct 10 15:09:04 crc kubenswrapper[4788]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 10 15:09:04 crc kubenswrapper[4788]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 10 15:09:04 crc kubenswrapper[4788]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + sleep 0.5 Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + sleep 0.5 Oct 10 15:09:04 crc kubenswrapper[4788]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 10 15:09:04 crc kubenswrapper[4788]: + cleanup_ovsdb_server_semaphore Oct 10 15:09:04 crc kubenswrapper[4788]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 10 15:09:04 crc kubenswrapper[4788]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 10 15:09:04 crc kubenswrapper[4788]: > pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" containerID="cri-o://310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.801915 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" containerID="cri-o://310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" gracePeriod=29 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.896664 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" containerID="cri-o://559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" gracePeriod=29 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923003 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-cmdqc"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923488 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b927859-0238-46f5-90aa-4afa094c434e","Type":"ContainerDied","Data":"61dd4ebbf802647ebbc1de9086c92e02600921dcbef187e7ed7f85e677087139"} Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923515 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-sbnwc"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923558 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-cmdqc"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923575 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7m7sz" event={"ID":"953774da-73b1-4f46-aa87-8deba0bcc077","Type":"ContainerDied","Data":"69a85b38a0e4154979d9d710277ca0a1c7a9d6c594026db5c938404d7dc3173b"} Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923590 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a9adb465-b660-4d2c-bae8-f0b8024761b8","Type":"ContainerDied","Data":"335000a0690e32f11a92cfd8dfc8fe055195d80c307ea823dc4da83c8237d39d"} Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923717 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-sbnwc"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923740 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n84gv" event={"ID":"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6","Type":"ContainerDied","Data":"94627d4d5603047c40017358c990871e5690c6c59e30cc3be499dae8b6bf6332"} Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.923756 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0bb0b-account-delete-lcsdj"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929534 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0bb0b-account-delete-lcsdj"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929582 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929597 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wkjwc"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929608 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-cnk7r"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929616 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-fr4km"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929653 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-cnk7r"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929667 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-fr4km"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929679 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-cpld9"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929694 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-cpld9"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929728 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-mxj9t"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929739 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-mxj9t"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929749 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2sjbr"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929759 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2sjbr"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929772 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-gxptq"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929781 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-gxptq"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929815 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929828 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5d565c7b56-z7s74"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929840 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929884 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-995b87f75-ghpz6"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929897 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929908 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.929919 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.930244 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerName="glance-log" containerID="cri-o://b5c371e53f6216ef8edfc3b5203d8b1eccf15ab2af6828b0ddf25a2c86e85ada" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.930434 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0bb0b-account-delete-lcsdj" Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.931164 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" podUID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" containerName="dnsmasq-dns" containerID="cri-o://a671f2494a1de22a55e001fea5624da5a53f37afb7bb091aee9590f2b7f67b6b" gracePeriod=10 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.931629 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-server" containerID="cri-o://3c7dd5fa50a9cff710f02fddbf4c3623aa2fc790f2bae6d263715c6a2a0be29a" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.931817 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5d565c7b56-z7s74" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-log" containerID="cri-o://f98704522449389d2af1c513f520e6f82ebedfbe293c1adf7038d2875138c32a" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.931994 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerName="cinder-scheduler" containerID="cri-o://82415a1ab836383b16f8eb7bf4bf4292b585bf668709c236d3369c260e68828a" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.934489 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-updater" containerID="cri-o://397c0d1e240027ba5e81338ba3e8dbb1354f4d23cae97d7d881c24cc8254d588" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.934630 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-995b87f75-ghpz6" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerName="neutron-api" containerID="cri-o://2861b006539119d480308e351b33bc584593a678f96dce7d7c0ab944bb039ad4" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.934731 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerName="glance-log" containerID="cri-o://ec39cee08096b21f5fb188e2f4be5f41b034a3a3eb4ccdb2d9860218975f6b3f" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.934826 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api-log" containerID="cri-o://a2c0776de729185f9cd35be86586a5c1c49c778ed6a5e56f595dcf61caf342d5" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935071 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-reaper" containerID="cri-o://3ffc6ad965d6e1a531e493d87c29df6c3edcf3b174067bab2eec1c073223fd6c" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935110 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-auditor" containerID="cri-o://bb676839b319bffb1dc7ed545f7044cc6b58cf9440fc39529f2f530eaababd0f" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935123 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-replicator" containerID="cri-o://292cc90a59ee5b84483d07d2af032969b7328c7cc913e39ba2daca604f9df39c" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935135 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-server" containerID="cri-o://d37f41271c5034f6009e04dd6c3c22abb922d7b658aa8853581bf17fef626476" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935198 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-updater" containerID="cri-o://4e4b1fea33830faff4707c7c32b35feb21cbafff038fad525b1d34bd1ddc3463" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935215 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-auditor" containerID="cri-o://da1d35d43d8e68922f14f6c617628576709461d2e0a8268d451f4cc8d694b90c" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935230 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-replicator" containerID="cri-o://7e17d3abb98cf36468ce697828f20b0a33f4bdac07da3b2f7a9793d696601d10" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935241 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-server" containerID="cri-o://642b90028a4eb24c71624969b7de73f0a0d672a8124912cb4fa0a195211e072e" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935275 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-auditor" containerID="cri-o://c68f1e8fc4064c0e04c7dd51691e00ff828f26e781caa02ae72f7ca08e509516" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935284 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-replicator" containerID="cri-o://bf79745568579090730afa53a1af2d611537375a823410f2aefdb5bb4ae1de4c" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.935311 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="rsync" containerID="cri-o://e1fa0e72188924318b11ba0840f43b627695d374bb14fe291577b87bd9109b71" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.937455 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerName="glance-httpd" containerID="cri-o://905b0fdae791afc90732669e1fd50c89e576f7fe20ab044f3b24b621f2e22853" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.937841 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerName="openstack-network-exporter" containerID="cri-o://f363963ef14d25b3ce98764062c7aee163273743e8ae4d349c9bc2b1bd5df94f" gracePeriod=300 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.937865 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="swift-recon-cron" containerID="cri-o://ea921c62533da26ac37b94fc9c80bfc74b41e75cddc9adab1f4225855ea56a7d" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.937937 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-expirer" containerID="cri-o://21748426974e63b0540c88af385f3f29ae489df2ec5f695b422d31559e86f185" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.938694 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerName="glance-httpd" containerID="cri-o://10a6407ac6ed46784d87d17ac1f81a2239476dc6186c82ebe7235c2468e4f18c" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.938723 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-995b87f75-ghpz6" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerName="neutron-httpd" containerID="cri-o://5699636a0ee4d24f0612d524e20711c956b0d7f42b60edadc77e819e598ab333" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.938780 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api" containerID="cri-o://dd729b335f5887ad2a47d9138d9341e607bccc05df0bd17180c3c60156b9f713" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.938822 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerName="probe" containerID="cri-o://737b4bebc0138dcddcb915236eb4ef30c70a15399b522e7c7159d53fa58d0ee4" gracePeriod=30 Oct 10 15:09:04 crc kubenswrapper[4788]: I1010 15:09:04.938833 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5d565c7b56-z7s74" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-api" containerID="cri-o://e39944b79b0c075079e9a288c1893a4cdbd681cc744af6cff568e6a8933d1c22" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.015501 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnlg9\" (UniqueName: \"kubernetes.io/projected/92ccaefc-72b3-4e2f-90c0-394240ed2ab2-kube-api-access-jnlg9\") pod \"novacell0bb0b-account-delete-lcsdj\" (UID: \"92ccaefc-72b3-4e2f-90c0-394240ed2ab2\") " pod="openstack/novacell0bb0b-account-delete-lcsdj" Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.021228 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-jmlx2"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.035471 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.179:8776/healthcheck\": read tcp 10.217.0.2:45572->10.217.0.179:8776: read: connection reset by peer" Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.051461 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-jmlx2"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.077304 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement4ba5-account-delete-ljq8m"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.093031 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4ba5-account-create-jhhcp"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.132714 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="ovsdbserver-sb" containerID="cri-o://5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d" gracePeriod=299 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.139048 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnlg9\" (UniqueName: \"kubernetes.io/projected/92ccaefc-72b3-4e2f-90c0-394240ed2ab2-kube-api-access-jnlg9\") pod \"novacell0bb0b-account-delete-lcsdj\" (UID: \"92ccaefc-72b3-4e2f-90c0-394240ed2ab2\") " pod="openstack/novacell0bb0b-account-delete-lcsdj" Oct 10 15:09:05 crc kubenswrapper[4788]: E1010 15:09:05.143958 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 10 15:09:05 crc kubenswrapper[4788]: E1010 15:09:05.144038 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data podName:dcf4b178-75ee-47fb-81e4-e6b277eb1790 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:07.144018645 +0000 UTC m=+1449.593734193 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data") pod "rabbitmq-server-0" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790") : configmap "rabbitmq-config-data" not found Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.189904 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4ba5-account-create-jhhcp"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.229261 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnlg9\" (UniqueName: \"kubernetes.io/projected/92ccaefc-72b3-4e2f-90c0-394240ed2ab2-kube-api-access-jnlg9\") pod \"novacell0bb0b-account-delete-lcsdj\" (UID: \"92ccaefc-72b3-4e2f-90c0-394240ed2ab2\") " pod="openstack/novacell0bb0b-account-delete-lcsdj" Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.343306 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerName="ovsdbserver-nb" containerID="cri-o://507ce22908f8278601f3c702001bc28ce6a57e91e11224453c2ba8d3538b36de" gracePeriod=300 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.423278 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.505932 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.506398 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-log" containerID="cri-o://5bc7218d76c149b7925383f9d5d6ab8180d2bb5b1b399cf25e15bef72a048be8" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.507026 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-metadata" containerID="cri-o://fab44d16043a2503f608627193df1f5950c0f9017dc038519c69453dcac75e2f" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.532917 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerName="rabbitmq" containerID="cri-o://c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251" gracePeriod=604800 Oct 10 15:09:05 crc kubenswrapper[4788]: E1010 15:09:05.577677 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:05 crc kubenswrapper[4788]: E1010 15:09:05.578126 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data podName:a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c nodeName:}" failed. No retries permitted until 2025-10-10 15:09:07.578110083 +0000 UTC m=+1450.027825631 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data") pod "rabbitmq-cell1-server-0" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c") : configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:05 crc kubenswrapper[4788]: E1010 15:09:05.607823 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedada8e6_c183_4ef8_bcd5_f5c617e7615e.slice/crio-3ffc6ad965d6e1a531e493d87c29df6c3edcf3b174067bab2eec1c073223fd6c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedada8e6_c183_4ef8_bcd5_f5c617e7615e.slice/crio-conmon-bf79745568579090730afa53a1af2d611537375a823410f2aefdb5bb4ae1de4c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedada8e6_c183_4ef8_bcd5_f5c617e7615e.slice/crio-conmon-e1fa0e72188924318b11ba0840f43b627695d374bb14fe291577b87bd9109b71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf57a61b_7d0e_4d49_b35b_d3059b5dd2ed.slice/crio-a671f2494a1de22a55e001fea5624da5a53f37afb7bb091aee9590f2b7f67b6b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b927859_0238_46f5_90aa_4afa094c434e.slice/crio-61dd4ebbf802647ebbc1de9086c92e02600921dcbef187e7ed7f85e677087139.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd84e4226_ac3f_4f6f_870b_c795e206854e.slice/crio-conmon-62068765dbbe9a789befff7b25adc8fe43b248f71c27fd2edbc0898452134000.scope\": RecentStats: unable to find data in memory cache]" Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.634788 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-87b66898f-smc95"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.635072 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-87b66898f-smc95" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-httpd" containerID="cri-o://fd696a35a1e5c7857218ceaedea76b07e06bb96d506d78970a1c5927133e9d90" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.636913 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-87b66898f-smc95" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-server" containerID="cri-o://87673bff375f480e6dff20a3a6b396e073584e22ecfa29abe8788960477c7c73" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.657377 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.657705 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-log" containerID="cri-o://79447d873480f8db4080543745a5c249b2cbec57b2c886298cf38288010480c5" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.662426 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-api" containerID="cri-o://598ec17f8706ac8a7613481f8dfaf31904983cb92d6010eefbf1ac47efcbb014" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.684818 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.704896 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-657c889cbb-d8d5h"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.705227 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerName="barbican-keystone-listener-log" containerID="cri-o://3c499468e5604cdfd26560a7f04ecbdd2ff30447627e047d1fc5e0e0d7e4cd3b" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.705726 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerName="barbican-keystone-listener" containerID="cri-o://705a07a6c50b14bb6069ed9b5c58612e9e5861f6620fb1c72d2da2cfd85e0e7b" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.724215 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-bc8db69bb-s2ddx"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.724709 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-bc8db69bb-s2ddx" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api-log" containerID="cri-o://441c3dc286f891485385a909d54daf00c136c609bb2e0e40814c55b244268d27" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.725574 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-bc8db69bb-s2ddx" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api" containerID="cri-o://f1acd019454438f9f318884feab766dfda302a7f2270489b71e0fc3660adf204" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.737009 4788 generic.go:334] "Generic (PLEG): container finished" podID="d84e4226-ac3f-4f6f-870b-c795e206854e" containerID="62068765dbbe9a789befff7b25adc8fe43b248f71c27fd2edbc0898452134000" exitCode=137 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.753128 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5b795d78cf-6sr7n"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.753701 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5b795d78cf-6sr7n" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" containerName="barbican-worker-log" containerID="cri-o://3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.753810 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-5b795d78cf-6sr7n" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" containerName="barbican-worker" containerID="cri-o://b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee" gracePeriod=30 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.770305 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.780007 4788 generic.go:334] "Generic (PLEG): container finished" podID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerID="f98704522449389d2af1c513f520e6f82ebedfbe293c1adf7038d2875138c32a" exitCode=143 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.780746 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d565c7b56-z7s74" event={"ID":"67000f4b-b955-44d7-ad37-3d552d8efd88","Type":"ContainerDied","Data":"f98704522449389d2af1c513f520e6f82ebedfbe293c1adf7038d2875138c32a"} Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.788184 4788 generic.go:334] "Generic (PLEG): container finished" podID="6627a65b-765d-430c-99f5-eddd285d4289" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" exitCode=0 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.788285 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-b986-account-create-hr255"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.788319 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wpgvn" event={"ID":"6627a65b-765d-430c-99f5-eddd285d4289","Type":"ContainerDied","Data":"310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92"} Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.798513 4788 generic.go:334] "Generic (PLEG): container finished" podID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerID="5699636a0ee4d24f0612d524e20711c956b0d7f42b60edadc77e819e598ab333" exitCode=0 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.798666 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-995b87f75-ghpz6" event={"ID":"e96b96e3-982f-4e27-97a6-4e076fc8ec40","Type":"ContainerDied","Data":"5699636a0ee4d24f0612d524e20711c956b0d7f42b60edadc77e819e598ab333"} Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.807189 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-79446"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.807947 4788 generic.go:334] "Generic (PLEG): container finished" podID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" containerID="a671f2494a1de22a55e001fea5624da5a53f37afb7bb091aee9590f2b7f67b6b" exitCode=0 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.808094 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" event={"ID":"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed","Type":"ContainerDied","Data":"a671f2494a1de22a55e001fea5624da5a53f37afb7bb091aee9590f2b7f67b6b"} Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.812438 4788 generic.go:334] "Generic (PLEG): container finished" podID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerID="ec39cee08096b21f5fb188e2f4be5f41b034a3a3eb4ccdb2d9860218975f6b3f" exitCode=143 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.812580 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"df85b9e1-2830-40c3-bbea-555bcddf5545","Type":"ContainerDied","Data":"ec39cee08096b21f5fb188e2f4be5f41b034a3a3eb4ccdb2d9860218975f6b3f"} Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.815747 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-b986-account-create-hr255"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.822778 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8951c1b9-28fd-449d-ade9-a7b7bfc3af6e/ovsdbserver-nb/0.log" Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.822863 4788 generic.go:334] "Generic (PLEG): container finished" podID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerID="f363963ef14d25b3ce98764062c7aee163273743e8ae4d349c9bc2b1bd5df94f" exitCode=2 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.822887 4788 generic.go:334] "Generic (PLEG): container finished" podID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerID="507ce22908f8278601f3c702001bc28ce6a57e91e11224453c2ba8d3538b36de" exitCode=143 Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.822967 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e","Type":"ContainerDied","Data":"f363963ef14d25b3ce98764062c7aee163273743e8ae4d349c9bc2b1bd5df94f"} Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.823005 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e","Type":"ContainerDied","Data":"507ce22908f8278601f3c702001bc28ce6a57e91e11224453c2ba8d3538b36de"} Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.841527 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-79446"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.857715 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderb986-account-delete-tvgzp"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.864739 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance9145-account-delete-kblv8" Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.866619 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-gxtng"] Oct 10 15:09:05 crc kubenswrapper[4788]: I1010 15:09:05.877196 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-gxtng"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.894173 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronab26-account-delete-g2dnr" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.907094 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-9145-account-create-vxxpk"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911878 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="e1fa0e72188924318b11ba0840f43b627695d374bb14fe291577b87bd9109b71" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911902 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="21748426974e63b0540c88af385f3f29ae489df2ec5f695b422d31559e86f185" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911916 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="397c0d1e240027ba5e81338ba3e8dbb1354f4d23cae97d7d881c24cc8254d588" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911923 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="bb676839b319bffb1dc7ed545f7044cc6b58cf9440fc39529f2f530eaababd0f" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911930 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="292cc90a59ee5b84483d07d2af032969b7328c7cc913e39ba2daca604f9df39c" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911937 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="4e4b1fea33830faff4707c7c32b35feb21cbafff038fad525b1d34bd1ddc3463" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911944 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="da1d35d43d8e68922f14f6c617628576709461d2e0a8268d451f4cc8d694b90c" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911952 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="7e17d3abb98cf36468ce697828f20b0a33f4bdac07da3b2f7a9793d696601d10" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911959 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="3ffc6ad965d6e1a531e493d87c29df6c3edcf3b174067bab2eec1c073223fd6c" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911966 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="c68f1e8fc4064c0e04c7dd51691e00ff828f26e781caa02ae72f7ca08e509516" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911977 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="bf79745568579090730afa53a1af2d611537375a823410f2aefdb5bb4ae1de4c" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.911927 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"e1fa0e72188924318b11ba0840f43b627695d374bb14fe291577b87bd9109b71"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912058 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"21748426974e63b0540c88af385f3f29ae489df2ec5f695b422d31559e86f185"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912070 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"397c0d1e240027ba5e81338ba3e8dbb1354f4d23cae97d7d881c24cc8254d588"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912080 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"bb676839b319bffb1dc7ed545f7044cc6b58cf9440fc39529f2f530eaababd0f"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912092 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"292cc90a59ee5b84483d07d2af032969b7328c7cc913e39ba2daca604f9df39c"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912101 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"4e4b1fea33830faff4707c7c32b35feb21cbafff038fad525b1d34bd1ddc3463"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912109 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"da1d35d43d8e68922f14f6c617628576709461d2e0a8268d451f4cc8d694b90c"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912117 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"7e17d3abb98cf36468ce697828f20b0a33f4bdac07da3b2f7a9793d696601d10"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912126 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"3ffc6ad965d6e1a531e493d87c29df6c3edcf3b174067bab2eec1c073223fd6c"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912134 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"c68f1e8fc4064c0e04c7dd51691e00ff828f26e781caa02ae72f7ca08e509516"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.912157 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"bf79745568579090730afa53a1af2d611537375a823410f2aefdb5bb4ae1de4c"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.920466 4788 generic.go:334] "Generic (PLEG): container finished" podID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerID="b5c371e53f6216ef8edfc3b5203d8b1eccf15ab2af6828b0ddf25a2c86e85ada" exitCode=143 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.920590 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2166478-a6f0-421b-9ea7-17f60ee9605b","Type":"ContainerDied","Data":"b5c371e53f6216ef8edfc3b5203d8b1eccf15ab2af6828b0ddf25a2c86e85ada"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.925803 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-9145-account-create-vxxpk"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.940398 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerName="rabbitmq" containerID="cri-o://a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab" gracePeriod=604800 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.948781 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b927859-0238-46f5-90aa-4afa094c434e/ovsdbserver-sb/0.log" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.948823 4788 generic.go:334] "Generic (PLEG): container finished" podID="4b927859-0238-46f5-90aa-4afa094c434e" containerID="5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d" exitCode=143 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.948874 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b927859-0238-46f5-90aa-4afa094c434e","Type":"ContainerDied","Data":"5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.956733 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-cmdtx"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.961971 4788 generic.go:334] "Generic (PLEG): container finished" podID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerID="a2c0776de729185f9cd35be86586a5c1c49c778ed6a5e56f595dcf61caf342d5" exitCode=143 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.962016 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ab6c469a-3a29-4fdb-b0ce-5671db6645d2","Type":"ContainerDied","Data":"a2c0776de729185f9cd35be86586a5c1c49c778ed6a5e56f595dcf61caf342d5"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.967440 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance9145-account-delete-kblv8"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:05.978962 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" containerName="galera" containerID="cri-o://06b8e0b90057fe848ada9d1d4fb30735ed3b1c3365172a8e3399572d5e67a2a7" gracePeriod=30 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.013077 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-cmdtx"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.027280 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ab26-account-create-24bhk"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.033763 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronab26-account-delete-g2dnr"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.040930 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ab26-account-create-24bhk"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.052626 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi96b1-account-delete-pm4dv" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.054752 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-96b1-account-create-vkctd"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.064911 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0bb0b-account-delete-lcsdj" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.072416 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7m7sz_953774da-73b1-4f46-aa87-8deba0bcc077/openstack-network-exporter/0.log" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.072512 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.073764 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-wrbdn"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.107267 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-metrics-certs-tls-certs\") pod \"953774da-73b1-4f46-aa87-8deba0bcc077\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.107368 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovs-rundir\") pod \"953774da-73b1-4f46-aa87-8deba0bcc077\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.107391 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s656\" (UniqueName: \"kubernetes.io/projected/953774da-73b1-4f46-aa87-8deba0bcc077-kube-api-access-5s656\") pod \"953774da-73b1-4f46-aa87-8deba0bcc077\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.107451 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-combined-ca-bundle\") pod \"953774da-73b1-4f46-aa87-8deba0bcc077\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.107640 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovn-rundir\") pod \"953774da-73b1-4f46-aa87-8deba0bcc077\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.107751 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953774da-73b1-4f46-aa87-8deba0bcc077-config\") pod \"953774da-73b1-4f46-aa87-8deba0bcc077\" (UID: \"953774da-73b1-4f46-aa87-8deba0bcc077\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.111879 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "953774da-73b1-4f46-aa87-8deba0bcc077" (UID: "953774da-73b1-4f46-aa87-8deba0bcc077"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.116062 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "953774da-73b1-4f46-aa87-8deba0bcc077" (UID: "953774da-73b1-4f46-aa87-8deba0bcc077"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.116906 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/953774da-73b1-4f46-aa87-8deba0bcc077-config" (OuterVolumeSpecName: "config") pod "953774da-73b1-4f46-aa87-8deba0bcc077" (UID: "953774da-73b1-4f46-aa87-8deba0bcc077"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.135459 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/953774da-73b1-4f46-aa87-8deba0bcc077-kube-api-access-5s656" (OuterVolumeSpecName: "kube-api-access-5s656") pod "953774da-73b1-4f46-aa87-8deba0bcc077" (UID: "953774da-73b1-4f46-aa87-8deba0bcc077"). InnerVolumeSpecName "kube-api-access-5s656". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.138186 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-96b1-account-create-vkctd"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.149280 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-wrbdn"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.163394 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi96b1-account-delete-pm4dv"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.178648 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-bb0b-account-create-bnxjn"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.192152 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-bb0b-account-create-bnxjn"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.195462 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "953774da-73b1-4f46-aa87-8deba0bcc077" (UID: "953774da-73b1-4f46-aa87-8deba0bcc077"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.196258 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n84gv" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.223351 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953774da-73b1-4f46-aa87-8deba0bcc077-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.223397 4788 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.223409 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s656\" (UniqueName: \"kubernetes.io/projected/953774da-73b1-4f46-aa87-8deba0bcc077-kube-api-access-5s656\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.223420 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.223432 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/953774da-73b1-4f46-aa87-8deba0bcc077-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.266105 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.273830 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f6f73b5-459a-4490-bac4-426d8b3f2b9e" path="/var/lib/kubelet/pods/1f6f73b5-459a-4490-bac4-426d8b3f2b9e/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.277176 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35c81559-a473-4667-8e31-71a388aa4cdf" path="/var/lib/kubelet/pods/35c81559-a473-4667-8e31-71a388aa4cdf/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.277864 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d" path="/var/lib/kubelet/pods/5b98f8e7-0f9c-44f7-9b8b-0a5638c0875d/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.278543 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8" path="/var/lib/kubelet/pods/5e68248c-def0-4dcc-a4ed-d8f9bd8cc5c8/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.282282 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ecb15a9-ac38-4eec-93f2-2069e6257ac0" path="/var/lib/kubelet/pods/5ecb15a9-ac38-4eec-93f2-2069e6257ac0/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.283238 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63cf2515-0dd5-4f1f-b70e-08a095284b53" path="/var/lib/kubelet/pods/63cf2515-0dd5-4f1f-b70e-08a095284b53/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.284079 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="679c6797-8be3-49ee-ada2-c2ea883db467" path="/var/lib/kubelet/pods/679c6797-8be3-49ee-ada2-c2ea883db467/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.285158 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7932fad1-fc89-4cde-86d4-210a1121424d" path="/var/lib/kubelet/pods/7932fad1-fc89-4cde-86d4-210a1121424d/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.287370 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02" path="/var/lib/kubelet/pods/a1dcaa1e-5539-4dc6-a1dc-a4cfb5637e02/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.288981 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aafa0506-d825-478a-9dc8-79f11ad28017" path="/var/lib/kubelet/pods/aafa0506-d825-478a-9dc8-79f11ad28017/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.290364 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0efd45c-b649-47ad-a008-349c8380a27a" path="/var/lib/kubelet/pods/b0efd45c-b649-47ad-a008-349c8380a27a/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.294470 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9f621aa-8944-4699-b706-53c84c2e9fe5" path="/var/lib/kubelet/pods/b9f621aa-8944-4699-b706-53c84c2e9fe5/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.296034 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb647ff7-3bd1-4325-a6a6-916a6f134fea" path="/var/lib/kubelet/pods/bb647ff7-3bd1-4325-a6a6-916a6f134fea/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.297087 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c95690b5-a444-46f4-98b6-997531cad4e8" path="/var/lib/kubelet/pods/c95690b5-a444-46f4-98b6-997531cad4e8/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.299067 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d89473fa-b587-4035-a586-80f51fa65786" path="/var/lib/kubelet/pods/d89473fa-b587-4035-a586-80f51fa65786/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.300517 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d7fc66-5e71-4858-96ef-fad2fc9189f9" path="/var/lib/kubelet/pods/d8d7fc66-5e71-4858-96ef-fad2fc9189f9/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.301214 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc982550-65dc-47de-b336-4e797e813403" path="/var/lib/kubelet/pods/dc982550-65dc-47de-b336-4e797e813403/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.302051 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2c43dc7-c35c-4406-9825-ae24803eba5e" path="/var/lib/kubelet/pods/e2c43dc7-c35c-4406-9825-ae24803eba5e/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.303491 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae89614-8508-4323-adf1-0692cedda9ca" path="/var/lib/kubelet/pods/eae89614-8508-4323-adf1-0692cedda9ca/volumes" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.314603 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "953774da-73b1-4f46-aa87-8deba0bcc077" (UID: "953774da-73b1-4f46-aa87-8deba0bcc077"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.320478 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0bb0b-account-delete-lcsdj"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.324727 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-w2642"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329472 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-combined-ca-bundle\") pod \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329618 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-config\") pod \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329694 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-swift-storage-0\") pod \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329726 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run\") pod \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329763 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-scripts\") pod \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329822 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-svc\") pod \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329894 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-log-ovn\") pod \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329929 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjk25\" (UniqueName: \"kubernetes.io/projected/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-kube-api-access-hjk25\") pod \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.329991 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-nb\") pod \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.330016 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-sb\") pod \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.330045 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run-ovn\") pod \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.330098 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrx79\" (UniqueName: \"kubernetes.io/projected/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-kube-api-access-xrx79\") pod \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\" (UID: \"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.330121 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-ovn-controller-tls-certs\") pod \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\" (UID: \"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6\") " Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.330661 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/953774da-73b1-4f46-aa87-8deba0bcc077-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.331002 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" (UID: "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.332411 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" (UID: "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.332611 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run" (OuterVolumeSpecName: "var-run") pod "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" (UID: "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.333648 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-scripts" (OuterVolumeSpecName: "scripts") pod "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" (UID: "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.345244 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-w2642"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.346679 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-kube-api-access-hjk25" (OuterVolumeSpecName: "kube-api-access-hjk25") pod "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" (UID: "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6"). InnerVolumeSpecName "kube-api-access-hjk25". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.366119 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dlwh4"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.375650 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dlwh4"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.381274 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-0a83-account-create-8k2jh"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.383813 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-kube-api-access-xrx79" (OuterVolumeSpecName: "kube-api-access-xrx79") pod "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" (UID: "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed"). InnerVolumeSpecName "kube-api-access-xrx79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.393263 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-0a83-account-create-8k2jh"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.398496 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.398737 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="c35da072-521f-4266-9600-688bf5b49c1c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d0bc3210873389e2972e8761b9e10f4be661565d640a62e62ffc9f730bc70876" gracePeriod=30 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.406046 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.406234 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="02360792-0295-42be-a94c-de0e7d2aa376" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e0ce65ec0009a40c1eb733283f7b65324c71d464e88882462d80a65bd07ae1af" gracePeriod=30 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.414060 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g5tcl"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.425946 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-g5tcl"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.432218 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d65h8"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.433123 4788 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.435354 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrx79\" (UniqueName: \"kubernetes.io/projected/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-kube-api-access-xrx79\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.435369 4788 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.435378 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.435385 4788 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.435393 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjk25\" (UniqueName: \"kubernetes.io/projected/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-kube-api-access-hjk25\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.438334 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-d65h8"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.444071 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.444388 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="cf293818-85fb-46dc-8cee-ba8eca827bb7" containerName="nova-cell0-conductor-conductor" containerID="cri-o://1ac5b4a4856f6b0a772902a9c4e92e64c3a65186c5c7b198248ff40fef09742d" gracePeriod=30 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.450983 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.451192 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4fad8d4a-3435-4647-b10d-51951c64e49a" containerName="nova-scheduler-scheduler" containerID="cri-o://edf59e020f4a2e0970646b96aab104bff00861373f9c92c494f167dd813d4c13" gracePeriod=30 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.467341 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" (UID: "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.468693 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderb986-account-delete-tvgzp"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.476191 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement4ba5-account-delete-ljq8m"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.486465 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican59ec-account-delete-nn85b"] Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.504351 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" (UID: "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.513130 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" (UID: "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.524896 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" (UID: "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.540577 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.540606 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.540617 4788 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.540625 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.587002 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-config" (OuterVolumeSpecName: "config") pod "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" (UID: "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.606650 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" (UID: "bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.613722 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" (UID: "64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.690029 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-87b66898f-smc95" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": dial tcp 10.217.0.166:8080: connect: connection refused" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.690117 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-87b66898f-smc95" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": dial tcp 10.217.0.166:8080: connect: connection refused" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.694985 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.695021 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.695032 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:06 crc kubenswrapper[4788]: E1010 15:09:06.803688 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 10 15:09:06 crc kubenswrapper[4788]: E1010 15:09:06.807079 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 10 15:09:06 crc kubenswrapper[4788]: E1010 15:09:06.809246 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 10 15:09:06 crc kubenswrapper[4788]: E1010 15:09:06.809292 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="ovn-northd" Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.981485 4788 generic.go:334] "Generic (PLEG): container finished" podID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" containerID="06b8e0b90057fe848ada9d1d4fb30735ed3b1c3365172a8e3399572d5e67a2a7" exitCode=0 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.981823 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a6c91be1-1707-4e13-a2ba-4d9cf48d9622","Type":"ContainerDied","Data":"06b8e0b90057fe848ada9d1d4fb30735ed3b1c3365172a8e3399572d5e67a2a7"} Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.984655 4788 generic.go:334] "Generic (PLEG): container finished" podID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerID="5bc7218d76c149b7925383f9d5d6ab8180d2bb5b1b399cf25e15bef72a048be8" exitCode=143 Oct 10 15:09:06 crc kubenswrapper[4788]: I1010 15:09:06.984720 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8","Type":"ContainerDied","Data":"5bc7218d76c149b7925383f9d5d6ab8180d2bb5b1b399cf25e15bef72a048be8"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.016533 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="d37f41271c5034f6009e04dd6c3c22abb922d7b658aa8853581bf17fef626476" exitCode=0 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.016581 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="642b90028a4eb24c71624969b7de73f0a0d672a8124912cb4fa0a195211e072e" exitCode=0 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.016588 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="3c7dd5fa50a9cff710f02fddbf4c3623aa2fc790f2bae6d263715c6a2a0be29a" exitCode=0 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.016631 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"d37f41271c5034f6009e04dd6c3c22abb922d7b658aa8853581bf17fef626476"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.016677 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"642b90028a4eb24c71624969b7de73f0a0d672a8124912cb4fa0a195211e072e"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.016687 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"3c7dd5fa50a9cff710f02fddbf4c3623aa2fc790f2bae6d263715c6a2a0be29a"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.019892 4788 generic.go:334] "Generic (PLEG): container finished" podID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerID="3c499468e5604cdfd26560a7f04ecbdd2ff30447627e047d1fc5e0e0d7e4cd3b" exitCode=143 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.019970 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" event={"ID":"085a2986-2c31-42f5-95c2-55f2414e5dfc","Type":"ContainerDied","Data":"3c499468e5604cdfd26560a7f04ecbdd2ff30447627e047d1fc5e0e0d7e4cd3b"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.026912 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" event={"ID":"bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed","Type":"ContainerDied","Data":"4efcd16f9cd65426df3b6309b3106a3ae25cf253d498d57c6f8be9e8fdda3e1c"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.026974 4788 scope.go:117] "RemoveContainer" containerID="a671f2494a1de22a55e001fea5624da5a53f37afb7bb091aee9590f2b7f67b6b" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.027134 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-wkjwc" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.034367 4788 generic.go:334] "Generic (PLEG): container finished" podID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerID="737b4bebc0138dcddcb915236eb4ef30c70a15399b522e7c7159d53fa58d0ee4" exitCode=0 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.034393 4788 generic.go:334] "Generic (PLEG): container finished" podID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerID="82415a1ab836383b16f8eb7bf4bf4292b585bf668709c236d3369c260e68828a" exitCode=0 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.034430 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898","Type":"ContainerDied","Data":"737b4bebc0138dcddcb915236eb4ef30c70a15399b522e7c7159d53fa58d0ee4"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.034452 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898","Type":"ContainerDied","Data":"82415a1ab836383b16f8eb7bf4bf4292b585bf668709c236d3369c260e68828a"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.038335 4788 generic.go:334] "Generic (PLEG): container finished" podID="7b604164-1660-4f3a-929b-cf6f59b66823" containerID="441c3dc286f891485385a909d54daf00c136c609bb2e0e40814c55b244268d27" exitCode=143 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.038378 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bc8db69bb-s2ddx" event={"ID":"7b604164-1660-4f3a-929b-cf6f59b66823","Type":"ContainerDied","Data":"441c3dc286f891485385a909d54daf00c136c609bb2e0e40814c55b244268d27"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.040038 4788 generic.go:334] "Generic (PLEG): container finished" podID="7810b9ee-1996-4535-bcac-5ffa8781295b" containerID="ae99f44876fccdb0bd4b1bbbc88a3459a502c4174dd932d08139e04e243eaf54" exitCode=0 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.040083 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb986-account-delete-tvgzp" event={"ID":"7810b9ee-1996-4535-bcac-5ffa8781295b","Type":"ContainerDied","Data":"ae99f44876fccdb0bd4b1bbbc88a3459a502c4174dd932d08139e04e243eaf54"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.040099 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb986-account-delete-tvgzp" event={"ID":"7810b9ee-1996-4535-bcac-5ffa8781295b","Type":"ContainerStarted","Data":"ed25769b4ff44d6002772563f03160fb6ef70b7bb07dc0cbca5d4f12486dec5e"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.057576 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican59ec-account-delete-nn85b" event={"ID":"5bfee301-605c-4fb8-992e-8911bfc7206f","Type":"ContainerStarted","Data":"bcac97c6393afdbb4439ab0c69c8425eca9cd3014bf3bb21a9bd851f12c6f668"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.081283 4788 generic.go:334] "Generic (PLEG): container finished" podID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerID="79447d873480f8db4080543745a5c249b2cbec57b2c886298cf38288010480c5" exitCode=143 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.081382 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b7f2c05-93d2-4139-95ea-c36bffcecdb8","Type":"ContainerDied","Data":"79447d873480f8db4080543745a5c249b2cbec57b2c886298cf38288010480c5"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.093828 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7m7sz" event={"ID":"953774da-73b1-4f46-aa87-8deba0bcc077","Type":"ContainerDied","Data":"6fe04936c82a41cfa98096d1c0aafacbbe3c2c071e21ef8ff5324af257d2fe60"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.093935 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7m7sz" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.106599 4788 generic.go:334] "Generic (PLEG): container finished" podID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerID="87673bff375f480e6dff20a3a6b396e073584e22ecfa29abe8788960477c7c73" exitCode=0 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.106631 4788 generic.go:334] "Generic (PLEG): container finished" podID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerID="fd696a35a1e5c7857218ceaedea76b07e06bb96d506d78970a1c5927133e9d90" exitCode=0 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.106681 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-87b66898f-smc95" event={"ID":"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f","Type":"ContainerDied","Data":"87673bff375f480e6dff20a3a6b396e073584e22ecfa29abe8788960477c7c73"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.106709 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-87b66898f-smc95" event={"ID":"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f","Type":"ContainerDied","Data":"fd696a35a1e5c7857218ceaedea76b07e06bb96d506d78970a1c5927133e9d90"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.122912 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement4ba5-account-delete-ljq8m" event={"ID":"024ba580-feec-4b2a-a836-942126b022ae","Type":"ContainerStarted","Data":"1030458b8b3ff8477d57735857a468ea3a2e6bd973b1045fcf998e6c108d3b3f"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.122961 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement4ba5-account-delete-ljq8m" event={"ID":"024ba580-feec-4b2a-a836-942126b022ae","Type":"ContainerStarted","Data":"36547aac8d71cdd733b2821304696a4dc1cbd3fe7b3fa455615fa933014afbe4"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.153829 4788 generic.go:334] "Generic (PLEG): container finished" podID="505a880b-6b48-498d-aad2-b30f20d089b6" containerID="3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950" exitCode=143 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.153913 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b795d78cf-6sr7n" event={"ID":"505a880b-6b48-498d-aad2-b30f20d089b6","Type":"ContainerDied","Data":"3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.164433 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-n84gv" event={"ID":"64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6","Type":"ContainerDied","Data":"f92de9a43d650c8d08e15516d369a71361f2a1c79f1946739aef031801b9812b"} Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.164574 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-n84gv" Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.234521 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.234656 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data podName:dcf4b178-75ee-47fb-81e4-e6b277eb1790 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:11.234630472 +0000 UTC m=+1453.684346020 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data") pod "rabbitmq-server-0" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790") : configmap "rabbitmq-config-data" not found Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.629265 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.630216 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="ceilometer-central-agent" containerID="cri-o://b2c6bf83b23d7eb9e36e2d8a2de6b0fcfeb53ed06aa0aaeb6182824c8e84405f" gracePeriod=30 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.630418 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="proxy-httpd" containerID="cri-o://cc438a9247888adf5836596dc8ce1d20b78dd2a805885fd538540aba1444f53e" gracePeriod=30 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.630486 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="sg-core" containerID="cri-o://0a1d2c1f952efc125bdddea9f7888d1cfbf3c24aa07f6a81eb81823168495ca6" gracePeriod=30 Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.630544 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="ceilometer-notification-agent" containerID="cri-o://6065428399b804c4b4b31556aa5431d788e52a568e70173b39bb05a529f66b04" gracePeriod=30 Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.645742 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.645842 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data podName:a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c nodeName:}" failed. No retries permitted until 2025-10-10 15:09:11.645820651 +0000 UTC m=+1454.095536199 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data") pod "rabbitmq-cell1-server-0" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c") : configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.661554 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0ce65ec0009a40c1eb733283f7b65324c71d464e88882462d80a65bd07ae1af" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.663842 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0ce65ec0009a40c1eb733283f7b65324c71d464e88882462d80a65bd07ae1af" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.666546 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.666760 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="6f13a166-2b9b-4679-877b-f4596b2d45ed" containerName="kube-state-metrics" containerID="cri-o://22d302e6d9d9eed7daebd5122f717fd594b1258085c469fab8d22ea7ffda2b6e" gracePeriod=30 Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.672460 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e0ce65ec0009a40c1eb733283f7b65324c71d464e88882462d80a65bd07ae1af" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.672532 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="02360792-0295-42be-a94c-de0e7d2aa376" containerName="nova-cell1-conductor-conductor" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.719293 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.719992 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d is running failed: container process not found" containerID="5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.732953 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d is running failed: container process not found" containerID="5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.735258 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d is running failed: container process not found" containerID="5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 10 15:09:07 crc kubenswrapper[4788]: E1010 15:09:07.735297 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="ovsdbserver-sb" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.740840 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-n84gv"] Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.747488 4788 scope.go:117] "RemoveContainer" containerID="70d6c0ca57d575e5a4f3a8e77dbaf0ffac929fea441a817e22ab138834ce2921" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.747808 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbjwq\" (UniqueName: \"kubernetes.io/projected/d84e4226-ac3f-4f6f-870b-c795e206854e-kube-api-access-zbjwq\") pod \"d84e4226-ac3f-4f6f-870b-c795e206854e\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.747884 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config-secret\") pod \"d84e4226-ac3f-4f6f-870b-c795e206854e\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.748019 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config\") pod \"d84e4226-ac3f-4f6f-870b-c795e206854e\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.748176 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-combined-ca-bundle\") pod \"d84e4226-ac3f-4f6f-870b-c795e206854e\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.748957 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-n84gv"] Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.752045 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.764408 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84e4226-ac3f-4f6f-870b-c795e206854e-kube-api-access-zbjwq" (OuterVolumeSpecName: "kube-api-access-zbjwq") pod "d84e4226-ac3f-4f6f-870b-c795e206854e" (UID: "d84e4226-ac3f-4f6f-870b-c795e206854e"). InnerVolumeSpecName "kube-api-access-zbjwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.806347 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8951c1b9-28fd-449d-ade9-a7b7bfc3af6e/ovsdbserver-nb/0.log" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.806439 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.807836 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.808378 4788 scope.go:117] "RemoveContainer" containerID="69a85b38a0e4154979d9d710277ca0a1c7a9d6c594026db5c938404d7dc3173b" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.809571 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b927859-0238-46f5-90aa-4afa094c434e/ovsdbserver-sb/0.log" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.809613 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.817238 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d84e4226-ac3f-4f6f-870b-c795e206854e" (UID: "d84e4226-ac3f-4f6f-870b-c795e206854e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.853028 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-7m7sz"] Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.857009 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d84e4226-ac3f-4f6f-870b-c795e206854e" (UID: "d84e4226-ac3f-4f6f-870b-c795e206854e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.895496 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config\") pod \"d84e4226-ac3f-4f6f-870b-c795e206854e\" (UID: \"d84e4226-ac3f-4f6f-870b-c795e206854e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.895896 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd69n\" (UniqueName: \"kubernetes.io/projected/4b927859-0238-46f5-90aa-4afa094c434e-kube-api-access-bd69n\") pod \"4b927859-0238-46f5-90aa-4afa094c434e\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.896721 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbjwq\" (UniqueName: \"kubernetes.io/projected/d84e4226-ac3f-4f6f-870b-c795e206854e-kube-api-access-zbjwq\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.896734 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:07 crc kubenswrapper[4788]: W1010 15:09:07.899251 4788 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/d84e4226-ac3f-4f6f-870b-c795e206854e/volumes/kubernetes.io~configmap/openstack-config Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.899270 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d84e4226-ac3f-4f6f-870b-c795e206854e" (UID: "d84e4226-ac3f-4f6f-870b-c795e206854e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.918203 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-7m7sz"] Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.942477 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b927859-0238-46f5-90aa-4afa094c434e-kube-api-access-bd69n" (OuterVolumeSpecName: "kube-api-access-bd69n") pod "4b927859-0238-46f5-90aa-4afa094c434e" (UID: "4b927859-0238-46f5-90aa-4afa094c434e"). InnerVolumeSpecName "kube-api-access-bd69n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.957487 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d84e4226-ac3f-4f6f-870b-c795e206854e" (UID: "d84e4226-ac3f-4f6f-870b-c795e206854e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.996302 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wkjwc"] Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997488 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-internal-tls-certs\") pod \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997543 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km8v9\" (UniqueName: \"kubernetes.io/projected/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-kube-api-access-km8v9\") pod \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997569 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmvpf\" (UniqueName: \"kubernetes.io/projected/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-kube-api-access-bmvpf\") pod \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997594 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"4b927859-0238-46f5-90aa-4afa094c434e\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997624 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b927859-0238-46f5-90aa-4afa094c434e-ovsdb-rundir\") pod \"4b927859-0238-46f5-90aa-4afa094c434e\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997644 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-public-tls-certs\") pod \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997680 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-run-httpd\") pod \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997708 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdb-rundir\") pod \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997746 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-scripts\") pod \"4b927859-0238-46f5-90aa-4afa094c434e\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.997780 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-config\") pod \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998000 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-config-data\") pod \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998027 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-combined-ca-bundle\") pod \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998047 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998075 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-metrics-certs-tls-certs\") pod \"4b927859-0238-46f5-90aa-4afa094c434e\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998103 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-scripts\") pod \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998123 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-etc-machine-id\") pod \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998153 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data-custom\") pod \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998182 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data\") pod \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998215 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-config\") pod \"4b927859-0238-46f5-90aa-4afa094c434e\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998233 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-ovsdbserver-sb-tls-certs\") pod \"4b927859-0238-46f5-90aa-4afa094c434e\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998262 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdbserver-nb-tls-certs\") pod \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998289 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-log-httpd\") pod \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998307 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-combined-ca-bundle\") pod \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998326 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-metrics-certs-tls-certs\") pod \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\" (UID: \"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998354 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-combined-ca-bundle\") pod \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998384 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-etc-swift\") pod \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998455 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpzpv\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-kube-api-access-gpzpv\") pod \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\" (UID: \"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998478 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-combined-ca-bundle\") pod \"4b927859-0238-46f5-90aa-4afa094c434e\" (UID: \"4b927859-0238-46f5-90aa-4afa094c434e\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998531 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-scripts\") pod \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\" (UID: \"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898\") " Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998943 4788 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998968 4788 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d84e4226-ac3f-4f6f-870b-c795e206854e-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:07 crc kubenswrapper[4788]: I1010 15:09:07.998983 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd69n\" (UniqueName: \"kubernetes.io/projected/4b927859-0238-46f5-90aa-4afa094c434e-kube-api-access-bd69n\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.001384 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" (UID: "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.005731 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-scripts" (OuterVolumeSpecName: "scripts") pod "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" (UID: "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.006324 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" (UID: "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.007029 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-config" (OuterVolumeSpecName: "config") pod "4b927859-0238-46f5-90aa-4afa094c434e" (UID: "4b927859-0238-46f5-90aa-4afa094c434e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.011838 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.020752 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-config" (OuterVolumeSpecName: "config") pod "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" (UID: "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.020966 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" (UID: "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.021220 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" (UID: "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.021317 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-scripts" (OuterVolumeSpecName: "scripts") pod "4b927859-0238-46f5-90aa-4afa094c434e" (UID: "4b927859-0238-46f5-90aa-4afa094c434e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.038391 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b927859-0238-46f5-90aa-4afa094c434e-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "4b927859-0238-46f5-90aa-4afa094c434e" (UID: "4b927859-0238-46f5-90aa-4afa094c434e"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.037722 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "4b927859-0238-46f5-90aa-4afa094c434e" (UID: "4b927859-0238-46f5-90aa-4afa094c434e"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.038691 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-kube-api-access-km8v9" (OuterVolumeSpecName: "kube-api-access-km8v9") pod "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" (UID: "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e"). InnerVolumeSpecName "kube-api-access-km8v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.065460 4788 scope.go:117] "RemoveContainer" containerID="94627d4d5603047c40017358c990871e5690c6c59e30cc3be499dae8b6bf6332" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.066131 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-kube-api-access-gpzpv" (OuterVolumeSpecName: "kube-api-access-gpzpv") pod "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" (UID: "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f"). InnerVolumeSpecName "kube-api-access-gpzpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.067551 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wkjwc"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.069122 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-kube-api-access-bmvpf" (OuterVolumeSpecName: "kube-api-access-bmvpf") pod "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" (UID: "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898"). InnerVolumeSpecName "kube-api-access-bmvpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.088337 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-scripts" (OuterVolumeSpecName: "scripts") pod "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" (UID: "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.100608 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-secrets\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.100682 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-galera-tls-certs\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.100755 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-operator-scripts\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.100783 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-combined-ca-bundle\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.100857 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kolla-config\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.100891 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-generated\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.100955 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-default\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.101009 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" (UID: "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.101659 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.102423 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.103120 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.103185 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cwh4\" (UniqueName: \"kubernetes.io/projected/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kube-api-access-8cwh4\") pod \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\" (UID: \"a6c91be1-1707-4e13-a2ba-4d9cf48d9622\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.103306 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.103479 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104521 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpzpv\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-kube-api-access-gpzpv\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104543 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104554 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km8v9\" (UniqueName: \"kubernetes.io/projected/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-kube-api-access-km8v9\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104566 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmvpf\" (UniqueName: \"kubernetes.io/projected/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-kube-api-access-bmvpf\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104579 4788 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104602 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104616 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4b927859-0238-46f5-90aa-4afa094c434e-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104629 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104639 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104651 4788 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104663 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104674 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104683 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104692 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104711 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104721 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104734 4788 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104745 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b927859-0238-46f5-90aa-4afa094c434e-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104756 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.104742 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" (UID: "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.123313 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" (UID: "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.124765 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kube-api-access-8cwh4" (OuterVolumeSpecName: "kube-api-access-8cwh4") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "kube-api-access-8cwh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.125099 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-secrets" (OuterVolumeSpecName: "secrets") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.191290 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.233266 4788 generic.go:334] "Generic (PLEG): container finished" podID="6f13a166-2b9b-4679-877b-f4596b2d45ed" containerID="22d302e6d9d9eed7daebd5122f717fd594b1258085c469fab8d22ea7ffda2b6e" exitCode=2 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.235306 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f13a166-2b9b-4679-877b-f4596b2d45ed","Type":"ContainerDied","Data":"22d302e6d9d9eed7daebd5122f717fd594b1258085c469fab8d22ea7ffda2b6e"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.240763 4788 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.240805 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.240815 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cwh4\" (UniqueName: \"kubernetes.io/projected/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-kube-api-access-8cwh4\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.240826 4788 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-secrets\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.240835 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.257298 4788 generic.go:334] "Generic (PLEG): container finished" podID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerID="705a07a6c50b14bb6069ed9b5c58612e9e5861f6620fb1c72d2da2cfd85e0e7b" exitCode=0 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.260507 4788 generic.go:334] "Generic (PLEG): container finished" podID="024ba580-feec-4b2a-a836-942126b022ae" containerID="1030458b8b3ff8477d57735857a468ea3a2e6bd973b1045fcf998e6c108d3b3f" exitCode=0 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.268184 4788 generic.go:334] "Generic (PLEG): container finished" podID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerID="cc438a9247888adf5836596dc8ce1d20b78dd2a805885fd538540aba1444f53e" exitCode=0 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.268227 4788 generic.go:334] "Generic (PLEG): container finished" podID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerID="0a1d2c1f952efc125bdddea9f7888d1cfbf3c24aa07f6a81eb81823168495ca6" exitCode=2 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.278489 4788 generic.go:334] "Generic (PLEG): container finished" podID="505a880b-6b48-498d-aad2-b30f20d089b6" containerID="b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee" exitCode=0 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.293825 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="244bd917-45fe-43e1-8386-029d12f5cb9a" path="/var/lib/kubelet/pods/244bd917-45fe-43e1-8386-029d12f5cb9a/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.294437 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab682b0-0d95-4150-8e87-88ad9a7d95a7" path="/var/lib/kubelet/pods/3ab682b0-0d95-4150-8e87-88ad9a7d95a7/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.295528 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" path="/var/lib/kubelet/pods/64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.296614 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64635416-9e49-4b2b-8b68-4ed756fbb05c" path="/var/lib/kubelet/pods/64635416-9e49-4b2b-8b68-4ed756fbb05c/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.297950 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="953774da-73b1-4f46-aa87-8deba0bcc077" path="/var/lib/kubelet/pods/953774da-73b1-4f46-aa87-8deba0bcc077/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.298846 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd483623-d8ec-4daa-9641-a25e3c92b0eb" path="/var/lib/kubelet/pods/bd483623-d8ec-4daa-9641-a25e3c92b0eb/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.299908 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" path="/var/lib/kubelet/pods/bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.301240 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d84e4226-ac3f-4f6f-870b-c795e206854e" path="/var/lib/kubelet/pods/d84e4226-ac3f-4f6f-870b-c795e206854e/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.301885 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd2cb897-fef8-4152-aacf-409fa2b13209" path="/var/lib/kubelet/pods/fd2cb897-fef8-4152-aacf-409fa2b13209/volumes" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.306182 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4b927859-0238-46f5-90aa-4afa094c434e/ovsdbserver-sb/0.log" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.306377 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.316824 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.324262 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_8951c1b9-28fd-449d-ade9-a7b7bfc3af6e/ovsdbserver-nb/0.log" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.325000 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.326946 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" (UID: "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.339390 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b927859-0238-46f5-90aa-4afa094c434e" (UID: "4b927859-0238-46f5-90aa-4afa094c434e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.339411 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" (UID: "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.342360 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.342392 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.342405 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.342439 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.345506 4788 generic.go:334] "Generic (PLEG): container finished" podID="5bfee301-605c-4fb8-992e-8911bfc7206f" containerID="fc75d105447e3e2c43eec1601b71054b73baff8f958dbf327e3eb0680d5cdba4" exitCode=0 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.352547 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.360062 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-87b66898f-smc95" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.362420 4788 generic.go:334] "Generic (PLEG): container finished" podID="02360792-0295-42be-a94c-de0e7d2aa376" containerID="e0ce65ec0009a40c1eb733283f7b65324c71d464e88882462d80a65bd07ae1af" exitCode=0 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.372461 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.373478 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.380599 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.390578 4788 generic.go:334] "Generic (PLEG): container finished" podID="c35da072-521f-4266-9600-688bf5b49c1c" containerID="d0bc3210873389e2972e8761b9e10f4be661565d640a62e62ffc9f730bc70876" exitCode=0 Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.393934 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.433902 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-5d565c7b56-z7s74" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-api" probeResult="failure" output="Get \"https://10.217.0.150:8778/\": read tcp 10.217.0.2:35908->10.217.0.150:8778: read: connection reset by peer" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.434324 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-5d565c7b56-z7s74" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.150:8778/\": read tcp 10.217.0.2:35912->10.217.0.150:8778: read: connection reset by peer" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.454550 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.454947 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.466318 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" (UID: "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.485301 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.501755 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "a6c91be1-1707-4e13-a2ba-4d9cf48d9622" (UID: "a6c91be1-1707-4e13-a2ba-4d9cf48d9622"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.556922 4788 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6c91be1-1707-4e13-a2ba-4d9cf48d9622-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.557457 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.557527 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.556130 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" (UID: "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.573401 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "4b927859-0238-46f5-90aa-4afa094c434e" (UID: "4b927859-0238-46f5-90aa-4afa094c434e"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.597041 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598736 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-sclcr"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598760 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-sclcr"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598862 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-59ec-account-create-dzhlt"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598879 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican59ec-account-delete-nn85b"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598893 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" event={"ID":"085a2986-2c31-42f5-95c2-55f2414e5dfc","Type":"ContainerDied","Data":"705a07a6c50b14bb6069ed9b5c58612e9e5861f6620fb1c72d2da2cfd85e0e7b"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598940 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-59ec-account-create-dzhlt"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598955 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement4ba5-account-delete-ljq8m" event={"ID":"024ba580-feec-4b2a-a836-942126b022ae","Type":"ContainerDied","Data":"1030458b8b3ff8477d57735857a468ea3a2e6bd973b1045fcf998e6c108d3b3f"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598969 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerDied","Data":"cc438a9247888adf5836596dc8ce1d20b78dd2a805885fd538540aba1444f53e"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598982 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerDied","Data":"0a1d2c1f952efc125bdddea9f7888d1cfbf3c24aa07f6a81eb81823168495ca6"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.598991 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b795d78cf-6sr7n" event={"ID":"505a880b-6b48-498d-aad2-b30f20d089b6","Type":"ContainerDied","Data":"b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599027 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5b795d78cf-6sr7n" event={"ID":"505a880b-6b48-498d-aad2-b30f20d089b6","Type":"ContainerDied","Data":"0731faf896e6f1b658e930bf659b512d62919816ddbf1dfae62248a5cc7acabb"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599038 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4b927859-0238-46f5-90aa-4afa094c434e","Type":"ContainerDied","Data":"4e9b28eae87eb831d3e9d85e3e537cebf371ca67643b93c002ab2145466e3187"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599052 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"8951c1b9-28fd-449d-ade9-a7b7bfc3af6e","Type":"ContainerDied","Data":"9c05966ceccac74c34fc128399f37a6a0c27ea686de2c2efe9f4b3d85242332d"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599067 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb986-account-delete-tvgzp" event={"ID":"7810b9ee-1996-4535-bcac-5ffa8781295b","Type":"ContainerDied","Data":"ed25769b4ff44d6002772563f03160fb6ef70b7bb07dc0cbca5d4f12486dec5e"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599123 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed25769b4ff44d6002772563f03160fb6ef70b7bb07dc0cbca5d4f12486dec5e" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599133 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican59ec-account-delete-nn85b" event={"ID":"5bfee301-605c-4fb8-992e-8911bfc7206f","Type":"ContainerDied","Data":"fc75d105447e3e2c43eec1601b71054b73baff8f958dbf327e3eb0680d5cdba4"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599225 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-87b66898f-smc95" event={"ID":"cd8f08dc-5aaa-4429-8f8d-190a36d5f38f","Type":"ContainerDied","Data":"048ed87aed0d749cfb7ffc903044db9c0bd9beac32a54479ee9f835b1b965c82"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599266 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"02360792-0295-42be-a94c-de0e7d2aa376","Type":"ContainerDied","Data":"e0ce65ec0009a40c1eb733283f7b65324c71d464e88882462d80a65bd07ae1af"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599284 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a6c91be1-1707-4e13-a2ba-4d9cf48d9622","Type":"ContainerDied","Data":"85878efb48481921ef0aac03f18cf613ad02ee52d29dc378e8bbffe20ab634f6"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599295 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898","Type":"ContainerDied","Data":"c55c29e788db344006016d530e0a778ed3eb9c13a119d66a5ef2f5b70e7705af"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.599305 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c35da072-521f-4266-9600-688bf5b49c1c","Type":"ContainerDied","Data":"d0bc3210873389e2972e8761b9e10f4be661565d640a62e62ffc9f730bc70876"} Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.604021 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" (UID: "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.609751 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderb986-account-delete-tvgzp" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.611132 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement4ba5-account-delete-ljq8m" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.621007 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "4b927859-0238-46f5-90aa-4afa094c434e" (UID: "4b927859-0238-46f5-90aa-4afa094c434e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.670655 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.670724 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.670736 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.670745 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b927859-0238-46f5-90aa-4afa094c434e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.671110 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data" (OuterVolumeSpecName: "config-data") pod "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" (UID: "05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.676235 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" (UID: "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.690449 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.201:3000/\": dial tcp 10.217.0.201:3000: connect: connection refused" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.722748 4788 scope.go:117] "RemoveContainer" containerID="b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.738628 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.771465 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-combined-ca-bundle\") pod \"505a880b-6b48-498d-aad2-b30f20d089b6\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.771493 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data\") pod \"505a880b-6b48-498d-aad2-b30f20d089b6\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.771526 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zbvb\" (UniqueName: \"kubernetes.io/projected/505a880b-6b48-498d-aad2-b30f20d089b6-kube-api-access-4zbvb\") pod \"505a880b-6b48-498d-aad2-b30f20d089b6\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.771738 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcgx6\" (UniqueName: \"kubernetes.io/projected/7810b9ee-1996-4535-bcac-5ffa8781295b-kube-api-access-fcgx6\") pod \"7810b9ee-1996-4535-bcac-5ffa8781295b\" (UID: \"7810b9ee-1996-4535-bcac-5ffa8781295b\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.771759 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a880b-6b48-498d-aad2-b30f20d089b6-logs\") pod \"505a880b-6b48-498d-aad2-b30f20d089b6\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.771774 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjpfm\" (UniqueName: \"kubernetes.io/projected/024ba580-feec-4b2a-a836-942126b022ae-kube-api-access-xjpfm\") pod \"024ba580-feec-4b2a-a836-942126b022ae\" (UID: \"024ba580-feec-4b2a-a836-942126b022ae\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.771843 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data-custom\") pod \"505a880b-6b48-498d-aad2-b30f20d089b6\" (UID: \"505a880b-6b48-498d-aad2-b30f20d089b6\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.772360 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.772374 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.775011 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/505a880b-6b48-498d-aad2-b30f20d089b6-logs" (OuterVolumeSpecName: "logs") pod "505a880b-6b48-498d-aad2-b30f20d089b6" (UID: "505a880b-6b48-498d-aad2-b30f20d089b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.797660 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.805726 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.860489 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "505a880b-6b48-498d-aad2-b30f20d089b6" (UID: "505a880b-6b48-498d-aad2-b30f20d089b6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.865602 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7810b9ee-1996-4535-bcac-5ffa8781295b-kube-api-access-fcgx6" (OuterVolumeSpecName: "kube-api-access-fcgx6") pod "7810b9ee-1996-4535-bcac-5ffa8781295b" (UID: "7810b9ee-1996-4535-bcac-5ffa8781295b"). InnerVolumeSpecName "kube-api-access-fcgx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.874344 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-vencrypt-tls-certs\") pod \"c35da072-521f-4266-9600-688bf5b49c1c\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.885234 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronab26-account-delete-g2dnr"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.893196 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-nova-novncproxy-tls-certs\") pod \"c35da072-521f-4266-9600-688bf5b49c1c\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.893274 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-combined-ca-bundle\") pod \"c35da072-521f-4266-9600-688bf5b49c1c\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.893299 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-combined-ca-bundle\") pod \"6f13a166-2b9b-4679-877b-f4596b2d45ed\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.893385 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjfvm\" (UniqueName: \"kubernetes.io/projected/c35da072-521f-4266-9600-688bf5b49c1c-kube-api-access-pjfvm\") pod \"c35da072-521f-4266-9600-688bf5b49c1c\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.893407 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-config-data\") pod \"c35da072-521f-4266-9600-688bf5b49c1c\" (UID: \"c35da072-521f-4266-9600-688bf5b49c1c\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.893491 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-config-data" (OuterVolumeSpecName: "config-data") pod "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" (UID: "cd8f08dc-5aaa-4429-8f8d-190a36d5f38f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.894194 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcgx6\" (UniqueName: \"kubernetes.io/projected/7810b9ee-1996-4535-bcac-5ffa8781295b-kube-api-access-fcgx6\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.894242 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/505a880b-6b48-498d-aad2-b30f20d089b6-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.894256 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.894265 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: E1010 15:09:08.894740 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:08 crc kubenswrapper[4788]: E1010 15:09:08.895030 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.895891 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/024ba580-feec-4b2a-a836-942126b022ae-kube-api-access-xjpfm" (OuterVolumeSpecName: "kube-api-access-xjpfm") pod "024ba580-feec-4b2a-a836-942126b022ae" (UID: "024ba580-feec-4b2a-a836-942126b022ae"). InnerVolumeSpecName "kube-api-access-xjpfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: E1010 15:09:08.924988 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:08 crc kubenswrapper[4788]: E1010 15:09:08.931264 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:08 crc kubenswrapper[4788]: E1010 15:09:08.931319 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.931425 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/505a880b-6b48-498d-aad2-b30f20d089b6-kube-api-access-4zbvb" (OuterVolumeSpecName: "kube-api-access-4zbvb") pod "505a880b-6b48-498d-aad2-b30f20d089b6" (UID: "505a880b-6b48-498d-aad2-b30f20d089b6"). InnerVolumeSpecName "kube-api-access-4zbvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.931619 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c35da072-521f-4266-9600-688bf5b49c1c-kube-api-access-pjfvm" (OuterVolumeSpecName: "kube-api-access-pjfvm") pod "c35da072-521f-4266-9600-688bf5b49c1c" (UID: "c35da072-521f-4266-9600-688bf5b49c1c"). InnerVolumeSpecName "kube-api-access-pjfvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.932649 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" (UID: "8951c1b9-28fd-449d-ade9-a7b7bfc3af6e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:08 crc kubenswrapper[4788]: E1010 15:09:08.962405 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.996580 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-combined-ca-bundle\") pod \"085a2986-2c31-42f5-95c2-55f2414e5dfc\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.996864 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2kqn\" (UniqueName: \"kubernetes.io/projected/085a2986-2c31-42f5-95c2-55f2414e5dfc-kube-api-access-m2kqn\") pod \"085a2986-2c31-42f5-95c2-55f2414e5dfc\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.996986 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kml9j\" (UniqueName: \"kubernetes.io/projected/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-api-access-kml9j\") pod \"6f13a166-2b9b-4679-877b-f4596b2d45ed\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.997211 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-certs\") pod \"6f13a166-2b9b-4679-877b-f4596b2d45ed\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.997350 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data\") pod \"085a2986-2c31-42f5-95c2-55f2414e5dfc\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.997458 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-config\") pod \"6f13a166-2b9b-4679-877b-f4596b2d45ed\" (UID: \"6f13a166-2b9b-4679-877b-f4596b2d45ed\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.997588 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data-custom\") pod \"085a2986-2c31-42f5-95c2-55f2414e5dfc\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.997657 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085a2986-2c31-42f5-95c2-55f2414e5dfc-logs\") pod \"085a2986-2c31-42f5-95c2-55f2414e5dfc\" (UID: \"085a2986-2c31-42f5-95c2-55f2414e5dfc\") " Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.998237 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjfvm\" (UniqueName: \"kubernetes.io/projected/c35da072-521f-4266-9600-688bf5b49c1c-kube-api-access-pjfvm\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.999174 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjpfm\" (UniqueName: \"kubernetes.io/projected/024ba580-feec-4b2a-a836-942126b022ae-kube-api-access-xjpfm\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.999285 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zbvb\" (UniqueName: \"kubernetes.io/projected/505a880b-6b48-498d-aad2-b30f20d089b6-kube-api-access-4zbvb\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.999342 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:08 crc kubenswrapper[4788]: I1010 15:09:08.998717 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi96b1-account-delete-pm4dv"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.041595 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0bb0b-account-delete-lcsdj"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.145604 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/085a2986-2c31-42f5-95c2-55f2414e5dfc-logs" (OuterVolumeSpecName: "logs") pod "085a2986-2c31-42f5-95c2-55f2414e5dfc" (UID: "085a2986-2c31-42f5-95c2-55f2414e5dfc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.165550 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "085a2986-2c31-42f5-95c2-55f2414e5dfc" (UID: "085a2986-2c31-42f5-95c2-55f2414e5dfc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.179788 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance9145-account-delete-kblv8"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.180767 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-api-access-kml9j" (OuterVolumeSpecName: "kube-api-access-kml9j") pod "6f13a166-2b9b-4679-877b-f4596b2d45ed" (UID: "6f13a166-2b9b-4679-877b-f4596b2d45ed"). InnerVolumeSpecName "kube-api-access-kml9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.193637 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kml9j\" (UniqueName: \"kubernetes.io/projected/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-api-access-kml9j\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.193679 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.193690 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/085a2986-2c31-42f5-95c2-55f2414e5dfc-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.193837 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085a2986-2c31-42f5-95c2-55f2414e5dfc-kube-api-access-m2kqn" (OuterVolumeSpecName: "kube-api-access-m2kqn") pod "085a2986-2c31-42f5-95c2-55f2414e5dfc" (UID: "085a2986-2c31-42f5-95c2-55f2414e5dfc"). InnerVolumeSpecName "kube-api-access-m2kqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.200070 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.200190 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.200309 4788 scope.go:117] "RemoveContainer" containerID="3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.231308 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:57666->10.217.0.204:8775: read: connection reset by peer" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.231472 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:57662->10.217.0.204:8775: read: connection reset by peer" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.249568 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f13a166-2b9b-4679-877b-f4596b2d45ed" (UID: "6f13a166-2b9b-4679-877b-f4596b2d45ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.255507 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.264200 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "505a880b-6b48-498d-aad2-b30f20d089b6" (UID: "505a880b-6b48-498d-aad2-b30f20d089b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.274154 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.310342 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ac5b4a4856f6b0a772902a9c4e92e64c3a65186c5c7b198248ff40fef09742d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.328812 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.333016 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "6f13a166-2b9b-4679-877b-f4596b2d45ed" (UID: "6f13a166-2b9b-4679-877b-f4596b2d45ed"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.333179 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c35da072-521f-4266-9600-688bf5b49c1c" (UID: "c35da072-521f-4266-9600-688bf5b49c1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.333476 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ac5b4a4856f6b0a772902a9c4e92e64c3a65186c5c7b198248ff40fef09742d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.344528 4788 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.344572 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.344583 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2kqn\" (UniqueName: \"kubernetes.io/projected/085a2986-2c31-42f5-95c2-55f2414e5dfc-kube-api-access-m2kqn\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.344596 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.344605 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.345462 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-config-data" (OuterVolumeSpecName: "config-data") pod "c35da072-521f-4266-9600-688bf5b49c1c" (UID: "c35da072-521f-4266-9600-688bf5b49c1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.403050 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-c44p7"] Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.403148 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ac5b4a4856f6b0a772902a9c4e92e64c3a65186c5c7b198248ff40fef09742d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.403203 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="cf293818-85fb-46dc-8cee-ba8eca827bb7" containerName="nova-cell0-conductor-conductor" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.403913 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "085a2986-2c31-42f5-95c2-55f2414e5dfc" (UID: "085a2986-2c31-42f5-95c2-55f2414e5dfc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.404270 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "c35da072-521f-4266-9600-688bf5b49c1c" (UID: "c35da072-521f-4266-9600-688bf5b49c1c"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.408084 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bc8db69bb-s2ddx" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:40400->10.217.0.159:9311: read: connection reset by peer" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.408905 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bc8db69bb-s2ddx" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:40412->10.217.0.159:9311: read: connection reset by peer" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.429705 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data" (OuterVolumeSpecName: "config-data") pod "085a2986-2c31-42f5-95c2-55f2414e5dfc" (UID: "085a2986-2c31-42f5-95c2-55f2414e5dfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.437193 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi96b1-account-delete-pm4dv" event={"ID":"0d375458-0dc4-45a2-abf9-c7376510d2f6","Type":"ContainerStarted","Data":"f7417620e756c0bf0389f9a52443ea4e26c42c102ca08ae937e8e3232743307a"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.437952 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-c44p7"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.439750 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" event={"ID":"085a2986-2c31-42f5-95c2-55f2414e5dfc","Type":"ContainerDied","Data":"3ca552bdec7a2eb28d25fc5b5aacd090a4e5b6635f4e9de4433c0f42afa4322a"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.439833 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-657c889cbb-d8d5h" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.449065 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.449251 4788 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.449315 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.449367 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085a2986-2c31-42f5-95c2-55f2414e5dfc-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.450736 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="edf59e020f4a2e0970646b96aab104bff00861373f9c92c494f167dd813d4c13" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.451799 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data" (OuterVolumeSpecName: "config-data") pod "505a880b-6b48-498d-aad2-b30f20d089b6" (UID: "505a880b-6b48-498d-aad2-b30f20d089b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.453083 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-t5zxq"] Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.472698 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="edf59e020f4a2e0970646b96aab104bff00861373f9c92c494f167dd813d4c13" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.478515 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="edf59e020f4a2e0970646b96aab104bff00861373f9c92c494f167dd813d4c13" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.478575 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4fad8d4a-3435-4647-b10d-51951c64e49a" containerName="nova-scheduler-scheduler" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.480482 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "6f13a166-2b9b-4679-877b-f4596b2d45ed" (UID: "6f13a166-2b9b-4679-877b-f4596b2d45ed"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.481522 4788 generic.go:334] "Generic (PLEG): container finished" podID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerID="b2c6bf83b23d7eb9e36e2d8a2de6b0fcfeb53ed06aa0aaeb6182824c8e84405f" exitCode=0 Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.481600 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerDied","Data":"b2c6bf83b23d7eb9e36e2d8a2de6b0fcfeb53ed06aa0aaeb6182824c8e84405f"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.487998 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance9145-account-delete-kblv8" event={"ID":"42d4825c-6e5c-4483-ae09-657903e6acb0","Type":"ContainerStarted","Data":"dd80b547d77a987656b69fe0189950742a72e4c531b826b36706620344d598b4"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.502927 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronab26-account-delete-g2dnr" event={"ID":"88f8c7ed-d478-4bae-ae23-4bffcb16abf6","Type":"ContainerStarted","Data":"d8adf007843a2beec7317755d482e4c66746a4321f0e4281a315c76cdb5da46e"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.502996 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-t5zxq"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.521318 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.527944 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "c35da072-521f-4266-9600-688bf5b49c1c" (UID: "c35da072-521f-4266-9600-688bf5b49c1c"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.533836 4788 generic.go:334] "Generic (PLEG): container finished" podID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerID="dd729b335f5887ad2a47d9138d9341e607bccc05df0bd17180c3c60156b9f713" exitCode=0 Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.533973 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ab6c469a-3a29-4fdb-b0ce-5671db6645d2","Type":"ContainerDied","Data":"dd729b335f5887ad2a47d9138d9341e607bccc05df0bd17180c3c60156b9f713"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.536614 4788 generic.go:334] "Generic (PLEG): container finished" podID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerID="e39944b79b0c075079e9a288c1893a4cdbd681cc744af6cff568e6a8933d1c22" exitCode=0 Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.536686 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d565c7b56-z7s74" event={"ID":"67000f4b-b955-44d7-ad37-3d552d8efd88","Type":"ContainerDied","Data":"e39944b79b0c075079e9a288c1893a4cdbd681cc744af6cff568e6a8933d1c22"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.538974 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f13a166-2b9b-4679-877b-f4596b2d45ed","Type":"ContainerDied","Data":"93b5237c52064c0a31349ac3298588c7ce002ea184a528a2127e6fc9caff6e44"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.538987 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.543903 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.551843 4788 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f13a166-2b9b-4679-877b-f4596b2d45ed-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.551873 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/505a880b-6b48-498d-aad2-b30f20d089b6-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.551886 4788 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35da072-521f-4266-9600-688bf5b49c1c-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.573008 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7f68d7b48c-24qxb"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.573436 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-7f68d7b48c-24qxb" podUID="970e7515-bd36-45d9-befb-a0597cc72cbb" containerName="keystone-api" containerID="cri-o://ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115" gracePeriod=30 Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.583719 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone56b0-account-delete-hfwph"] Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.584844 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7810b9ee-1996-4535-bcac-5ffa8781295b" containerName="mariadb-account-delete" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.584862 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7810b9ee-1996-4535-bcac-5ffa8781295b" containerName="mariadb-account-delete" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.584910 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" containerName="mysql-bootstrap" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.584918 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" containerName="mysql-bootstrap" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.584927 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.584933 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.584947 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="ovsdbserver-sb" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.584954 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="ovsdbserver-sb" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.584962 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f13a166-2b9b-4679-877b-f4596b2d45ed" containerName="kube-state-metrics" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.584970 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f13a166-2b9b-4679-877b-f4596b2d45ed" containerName="kube-state-metrics" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.584978 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerName="barbican-keystone-listener-log" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.584984 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerName="barbican-keystone-listener-log" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.584995 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerName="probe" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.585002 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerName="probe" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.585017 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.585023 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.585035 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" containerName="barbican-worker" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.585042 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" containerName="barbican-worker" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.585057 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-httpd" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.585064 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-httpd" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.585077 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="953774da-73b1-4f46-aa87-8deba0bcc077" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.585084 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="953774da-73b1-4f46-aa87-8deba0bcc077" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.585095 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerName="cinder-scheduler" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.585101 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerName="cinder-scheduler" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.585112 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerName="barbican-keystone-listener" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.585118 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerName="barbican-keystone-listener" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.585129 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" containerName="ovn-controller" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590169 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" containerName="ovn-controller" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.590205 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="024ba580-feec-4b2a-a836-942126b022ae" containerName="mariadb-account-delete" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590215 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="024ba580-feec-4b2a-a836-942126b022ae" containerName="mariadb-account-delete" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.590244 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" containerName="galera" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590252 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" containerName="galera" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.590280 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" containerName="barbican-worker-log" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590287 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" containerName="barbican-worker-log" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.590297 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerName="ovsdbserver-nb" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590302 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerName="ovsdbserver-nb" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.590313 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35da072-521f-4266-9600-688bf5b49c1c" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590319 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35da072-521f-4266-9600-688bf5b49c1c" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.590330 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" containerName="dnsmasq-dns" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590336 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" containerName="dnsmasq-dns" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.590359 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" containerName="init" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590365 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" containerName="init" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.590381 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-server" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590387 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-server" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590759 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f13a166-2b9b-4679-877b-f4596b2d45ed" containerName="kube-state-metrics" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590775 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf57a61b-7d0e-4d49-b35b-d3059b5dd2ed" containerName="dnsmasq-dns" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590784 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerName="probe" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590796 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" containerName="galera" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590812 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerName="barbican-keystone-listener-log" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590823 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35da072-521f-4266-9600-688bf5b49c1c" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590833 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="024ba580-feec-4b2a-a836-942126b022ae" containerName="mariadb-account-delete" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590838 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="ovsdbserver-sb" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590848 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-server" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590855 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="953774da-73b1-4f46-aa87-8deba0bcc077" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590861 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" containerName="barbican-worker" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590870 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590879 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" containerName="barbican-keystone-listener" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590886 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" containerName="cinder-scheduler" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590895 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="64489fdb-6fb3-47ae-b3f0-c3e16c8de3a6" containerName="ovn-controller" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590903 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" containerName="barbican-worker-log" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590914 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b927859-0238-46f5-90aa-4afa094c434e" containerName="openstack-network-exporter" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590924 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" containerName="proxy-httpd" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590934 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7810b9ee-1996-4535-bcac-5ffa8781295b" containerName="mariadb-account-delete" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.590942 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" containerName="ovsdbserver-nb" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.593043 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.598435 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"02360792-0295-42be-a94c-de0e7d2aa376","Type":"ContainerDied","Data":"9f77918f6dbccdb5b5e3ebb8be7846a782d1131d10ba2ad042638d4259dee7bf"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.598489 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f77918f6dbccdb5b5e3ebb8be7846a782d1131d10ba2ad042638d4259dee7bf" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.604274 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement4ba5-account-delete-ljq8m" event={"ID":"024ba580-feec-4b2a-a836-942126b022ae","Type":"ContainerDied","Data":"36547aac8d71cdd733b2821304696a4dc1cbd3fe7b3fa455615fa933014afbe4"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.604440 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement4ba5-account-delete-ljq8m" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.624446 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0bb0b-account-delete-lcsdj" event={"ID":"92ccaefc-72b3-4e2f-90c0-394240ed2ab2","Type":"ContainerStarted","Data":"6da28bc2ea7d4101fea6c67bfcb3e698df4aa0c013b90d1e81ec7d3a016e1f77"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.625646 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.630898 4788 generic.go:334] "Generic (PLEG): container finished" podID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerID="fab44d16043a2503f608627193df1f5950c0f9017dc038519c69453dcac75e2f" exitCode=0 Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.631038 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8","Type":"ContainerDied","Data":"fab44d16043a2503f608627193df1f5950c0f9017dc038519c69453dcac75e2f"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.635021 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c35da072-521f-4266-9600-688bf5b49c1c","Type":"ContainerDied","Data":"bf9f470005274a391d3390ce8ec6e5159f552cd80b56c76a008259e229d503f8"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.635207 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.637849 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone56b0-account-delete-hfwph"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.644331 4788 generic.go:334] "Generic (PLEG): container finished" podID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerID="905b0fdae791afc90732669e1fd50c89e576f7fe20ab044f3b24b621f2e22853" exitCode=0 Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.644404 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2166478-a6f0-421b-9ea7-17f60ee9605b","Type":"ContainerDied","Data":"905b0fdae791afc90732669e1fd50c89e576f7fe20ab044f3b24b621f2e22853"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.648053 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5b795d78cf-6sr7n" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.649434 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-mml9f"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.655771 4788 generic.go:334] "Generic (PLEG): container finished" podID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerID="10a6407ac6ed46784d87d17ac1f81a2239476dc6186c82ebe7235c2468e4f18c" exitCode=0 Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.655891 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"df85b9e1-2830-40c3-bbea-555bcddf5545","Type":"ContainerDied","Data":"10a6407ac6ed46784d87d17ac1f81a2239476dc6186c82ebe7235c2468e4f18c"} Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.655910 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderb986-account-delete-tvgzp" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.656804 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="4075bf4d-01ed-4d7c-9fb8-576dd2866821" containerName="memcached" containerID="cri-o://2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9" gracePeriod=30 Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.683483 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-mml9f"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.693213 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone56b0-account-delete-hfwph"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.697325 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-56b0-account-create-wqxf2"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.706117 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-56b0-account-create-wqxf2"] Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.758816 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz5rc\" (UniqueName: \"kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc\") pod \"keystone56b0-account-delete-hfwph\" (UID: \"3aa6d9f2-ee07-4cff-9cdc-def5e94ede33\") " pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:09 crc kubenswrapper[4788]: I1010 15:09:09.861908 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz5rc\" (UniqueName: \"kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc\") pod \"keystone56b0-account-delete-hfwph\" (UID: \"3aa6d9f2-ee07-4cff-9cdc-def5e94ede33\") " pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.869424 4788 projected.go:194] Error preparing data for projected volume kube-api-access-pz5rc for pod openstack/keystone56b0-account-delete-hfwph: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 10 15:09:09 crc kubenswrapper[4788]: E1010 15:09:09.869512 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc podName:3aa6d9f2-ee07-4cff-9cdc-def5e94ede33 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:10.369490952 +0000 UTC m=+1452.819206500 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-pz5rc" (UniqueName: "kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc") pod "keystone56b0-account-delete-hfwph" (UID: "3aa6d9f2-ee07-4cff-9cdc-def5e94ede33") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.009952 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerName="galera" containerID="cri-o://36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74" gracePeriod=30 Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.212209 4788 scope.go:117] "RemoveContainer" containerID="b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee" Oct 10 15:09:10 crc kubenswrapper[4788]: E1010 15:09:10.231612 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee\": container with ID starting with b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee not found: ID does not exist" containerID="b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.231681 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee"} err="failed to get container status \"b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee\": rpc error: code = NotFound desc = could not find container \"b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee\": container with ID starting with b40569cb7e8ef5c5a8c72faa5ff8f392baa974b6bd030a65ed0542063a640cee not found: ID does not exist" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.231736 4788 scope.go:117] "RemoveContainer" containerID="3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950" Oct 10 15:09:10 crc kubenswrapper[4788]: E1010 15:09:10.237814 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950\": container with ID starting with 3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950 not found: ID does not exist" containerID="3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.237877 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950"} err="failed to get container status \"3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950\": rpc error: code = NotFound desc = could not find container \"3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950\": container with ID starting with 3591b28e18ba6c3422ad82ec20e86f5c92d91b4dd8f202214c122464cac13950 not found: ID does not exist" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.237903 4788 scope.go:117] "RemoveContainer" containerID="61dd4ebbf802647ebbc1de9086c92e02600921dcbef187e7ed7f85e677087139" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.248322 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898" path="/var/lib/kubelet/pods/05eb5c0f-901b-43d2-a6f9-ed7fe5ba1898/volumes" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.249237 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1063e9ab-7517-4a5e-8865-f7827c52245a" path="/var/lib/kubelet/pods/1063e9ab-7517-4a5e-8865-f7827c52245a/volumes" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.250431 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1586c9ff-8c4a-4946-b524-cf90d3d81db9" path="/var/lib/kubelet/pods/1586c9ff-8c4a-4946-b524-cf90d3d81db9/volumes" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.251482 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dc79ccd-3195-4403-a56a-01c7ff0984a9" path="/var/lib/kubelet/pods/5dc79ccd-3195-4403-a56a-01c7ff0984a9/volumes" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.252278 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6c91be1-1707-4e13-a2ba-4d9cf48d9622" path="/var/lib/kubelet/pods/a6c91be1-1707-4e13-a2ba-4d9cf48d9622/volumes" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.252808 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b098f0-5965-4ab3-a641-e301b2f599ed" path="/var/lib/kubelet/pods/b5b098f0-5965-4ab3-a641-e301b2f599ed/volumes" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.253717 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df37fb0e-106c-4fe4-8652-cad4d132d205" path="/var/lib/kubelet/pods/df37fb0e-106c-4fe4-8652-cad4d132d205/volumes" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.254198 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f12ed2f7-2d5e-4a57-bd17-59537de8065a" path="/var/lib/kubelet/pods/f12ed2f7-2d5e-4a57-bd17-59537de8065a/volumes" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.373882 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz5rc\" (UniqueName: \"kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc\") pod \"keystone56b0-account-delete-hfwph\" (UID: \"3aa6d9f2-ee07-4cff-9cdc-def5e94ede33\") " pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:10 crc kubenswrapper[4788]: E1010 15:09:10.379816 4788 projected.go:194] Error preparing data for projected volume kube-api-access-pz5rc for pod openstack/keystone56b0-account-delete-hfwph: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 10 15:09:10 crc kubenswrapper[4788]: E1010 15:09:10.379887 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc podName:3aa6d9f2-ee07-4cff-9cdc-def5e94ede33 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:11.379866896 +0000 UTC m=+1453.829582444 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-pz5rc" (UniqueName: "kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc") pod "keystone56b0-account-delete-hfwph" (UID: "3aa6d9f2-ee07-4cff-9cdc-def5e94ede33") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.401429 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.481831 4788 scope.go:117] "RemoveContainer" containerID="5c50c1d9beb989f8fe1c1fa6134f93f1c87999a80fd464b6d74e163f6043ec2d" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.543346 4788 scope.go:117] "RemoveContainer" containerID="f363963ef14d25b3ce98764062c7aee163273743e8ae4d349c9bc2b1bd5df94f" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.549485 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.568079 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.584239 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvh8t\" (UniqueName: \"kubernetes.io/projected/02360792-0295-42be-a94c-de0e7d2aa376-kube-api-access-wvh8t\") pod \"02360792-0295-42be-a94c-de0e7d2aa376\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.584909 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-combined-ca-bundle\") pod \"02360792-0295-42be-a94c-de0e7d2aa376\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.585126 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-config-data\") pod \"02360792-0295-42be-a94c-de0e7d2aa376\" (UID: \"02360792-0295-42be-a94c-de0e7d2aa376\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.604360 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02360792-0295-42be-a94c-de0e7d2aa376-kube-api-access-wvh8t" (OuterVolumeSpecName: "kube-api-access-wvh8t") pod "02360792-0295-42be-a94c-de0e7d2aa376" (UID: "02360792-0295-42be-a94c-de0e7d2aa376"). InnerVolumeSpecName "kube-api-access-wvh8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.681078 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.682025 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2166478-a6f0-421b-9ea7-17f60ee9605b","Type":"ContainerDied","Data":"3022642694ba4ddf812aab1db3f5d23fe1d48af8e85b4b87a5d823f6c2e80f19"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.687698 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-httpd-run\") pod \"e2166478-a6f0-421b-9ea7-17f60ee9605b\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.687766 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-scripts\") pod \"e2166478-a6f0-421b-9ea7-17f60ee9605b\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.687834 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-logs\") pod \"e2166478-a6f0-421b-9ea7-17f60ee9605b\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.687856 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-config-data\") pod \"e2166478-a6f0-421b-9ea7-17f60ee9605b\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.687872 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp6sp\" (UniqueName: \"kubernetes.io/projected/e2166478-a6f0-421b-9ea7-17f60ee9605b-kube-api-access-xp6sp\") pod \"e2166478-a6f0-421b-9ea7-17f60ee9605b\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.687950 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-combined-ca-bundle\") pod \"e2166478-a6f0-421b-9ea7-17f60ee9605b\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.688060 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"e2166478-a6f0-421b-9ea7-17f60ee9605b\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.688101 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-public-tls-certs\") pod \"e2166478-a6f0-421b-9ea7-17f60ee9605b\" (UID: \"e2166478-a6f0-421b-9ea7-17f60ee9605b\") " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.688492 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvh8t\" (UniqueName: \"kubernetes.io/projected/02360792-0295-42be-a94c-de0e7d2aa376-kube-api-access-wvh8t\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.689336 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e2166478-a6f0-421b-9ea7-17f60ee9605b" (UID: "e2166478-a6f0-421b-9ea7-17f60ee9605b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.691170 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-logs" (OuterVolumeSpecName: "logs") pod "e2166478-a6f0-421b-9ea7-17f60ee9605b" (UID: "e2166478-a6f0-421b-9ea7-17f60ee9605b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.694157 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.696519 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"df85b9e1-2830-40c3-bbea-555bcddf5545","Type":"ContainerDied","Data":"73b7decab65445051dfd83cb6ad434f9bf9dbe77af7ef1c02e9a1752cb60a81b"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.696557 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73b7decab65445051dfd83cb6ad434f9bf9dbe77af7ef1c02e9a1752cb60a81b" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.701067 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2166478-a6f0-421b-9ea7-17f60ee9605b-kube-api-access-xp6sp" (OuterVolumeSpecName: "kube-api-access-xp6sp") pod "e2166478-a6f0-421b-9ea7-17f60ee9605b" (UID: "e2166478-a6f0-421b-9ea7-17f60ee9605b"). InnerVolumeSpecName "kube-api-access-xp6sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.703048 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02360792-0295-42be-a94c-de0e7d2aa376" (UID: "02360792-0295-42be-a94c-de0e7d2aa376"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.703060 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8","Type":"ContainerDied","Data":"1ce4d467f7c9d92a86b58fa6239577d1a24246e94ca1434de7ba399a191e8d73"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.703171 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ce4d467f7c9d92a86b58fa6239577d1a24246e94ca1434de7ba399a191e8d73" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.705728 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-scripts" (OuterVolumeSpecName: "scripts") pod "e2166478-a6f0-421b-9ea7-17f60ee9605b" (UID: "e2166478-a6f0-421b-9ea7-17f60ee9605b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.706260 4788 generic.go:334] "Generic (PLEG): container finished" podID="cf293818-85fb-46dc-8cee-ba8eca827bb7" containerID="1ac5b4a4856f6b0a772902a9c4e92e64c3a65186c5c7b198248ff40fef09742d" exitCode=0 Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.706347 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cf293818-85fb-46dc-8cee-ba8eca827bb7","Type":"ContainerDied","Data":"1ac5b4a4856f6b0a772902a9c4e92e64c3a65186c5c7b198248ff40fef09742d"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.708724 4788 generic.go:334] "Generic (PLEG): container finished" podID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerID="598ec17f8706ac8a7613481f8dfaf31904983cb92d6010eefbf1ac47efcbb014" exitCode=0 Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.708734 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "e2166478-a6f0-421b-9ea7-17f60ee9605b" (UID: "e2166478-a6f0-421b-9ea7-17f60ee9605b"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.708794 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b7f2c05-93d2-4139-95ea-c36bffcecdb8","Type":"ContainerDied","Data":"598ec17f8706ac8a7613481f8dfaf31904983cb92d6010eefbf1ac47efcbb014"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.708837 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4b7f2c05-93d2-4139-95ea-c36bffcecdb8","Type":"ContainerDied","Data":"5a890daae46b4f28ec56c1ab8637d5a49495dfd1c24de91adea89c800024351a"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.708860 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a890daae46b4f28ec56c1ab8637d5a49495dfd1c24de91adea89c800024351a" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.711326 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5d565c7b56-z7s74" event={"ID":"67000f4b-b955-44d7-ad37-3d552d8efd88","Type":"ContainerDied","Data":"4548a22d7e0d810257f4a467ffef7507c6f03b2a25499966819f1ef5c4552182"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.711422 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4548a22d7e0d810257f4a467ffef7507c6f03b2a25499966819f1ef5c4552182" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.723244 4788 generic.go:334] "Generic (PLEG): container finished" podID="7b604164-1660-4f3a-929b-cf6f59b66823" containerID="f1acd019454438f9f318884feab766dfda302a7f2270489b71e0fc3660adf204" exitCode=0 Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.723318 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bc8db69bb-s2ddx" event={"ID":"7b604164-1660-4f3a-929b-cf6f59b66823","Type":"ContainerDied","Data":"f1acd019454438f9f318884feab766dfda302a7f2270489b71e0fc3660adf204"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.723351 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bc8db69bb-s2ddx" event={"ID":"7b604164-1660-4f3a-929b-cf6f59b66823","Type":"ContainerDied","Data":"7af9a6e24bee3d2c197114ed28318653e221a3c48158feaad3ed7eb762d58f09"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.723399 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7af9a6e24bee3d2c197114ed28318653e221a3c48158feaad3ed7eb762d58f09" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.735928 4788 generic.go:334] "Generic (PLEG): container finished" podID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerID="6065428399b804c4b4b31556aa5431d788e52a568e70173b39bb05a529f66b04" exitCode=0 Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.736002 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerDied","Data":"6065428399b804c4b4b31556aa5431d788e52a568e70173b39bb05a529f66b04"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.736047 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4aeeba9b-9894-4293-a673-7e88c6bc80f5","Type":"ContainerDied","Data":"9512de2e304f913ec19e124ad1197def2e433a95716f5eb2209fe64db2f17616"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.736062 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9512de2e304f913ec19e124ad1197def2e433a95716f5eb2209fe64db2f17616" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.740762 4788 generic.go:334] "Generic (PLEG): container finished" podID="4fad8d4a-3435-4647-b10d-51951c64e49a" containerID="edf59e020f4a2e0970646b96aab104bff00861373f9c92c494f167dd813d4c13" exitCode=0 Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.740963 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4fad8d4a-3435-4647-b10d-51951c64e49a","Type":"ContainerDied","Data":"edf59e020f4a2e0970646b96aab104bff00861373f9c92c494f167dd813d4c13"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.741048 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4fad8d4a-3435-4647-b10d-51951c64e49a","Type":"ContainerDied","Data":"48a796b250ee9e5c80cca00d0a66e4bafe6672c6c42fe572bb77dfc8eadfcfa4"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.741106 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48a796b250ee9e5c80cca00d0a66e4bafe6672c6c42fe572bb77dfc8eadfcfa4" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.745600 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ab6c469a-3a29-4fdb-b0ce-5671db6645d2","Type":"ContainerDied","Data":"a3cb17cdb6ba13e494afad0db5282252dc94d4f28170a0f1d6259e36b5f20949"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.745731 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3cb17cdb6ba13e494afad0db5282252dc94d4f28170a0f1d6259e36b5f20949" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.748601 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.749032 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican59ec-account-delete-nn85b" event={"ID":"5bfee301-605c-4fb8-992e-8911bfc7206f","Type":"ContainerDied","Data":"bcac97c6393afdbb4439ab0c69c8425eca9cd3014bf3bb21a9bd851f12c6f668"} Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.749309 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcac97c6393afdbb4439ab0c69c8425eca9cd3014bf3bb21a9bd851f12c6f668" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.751189 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-config-data" (OuterVolumeSpecName: "config-data") pod "02360792-0295-42be-a94c-de0e7d2aa376" (UID: "02360792-0295-42be-a94c-de0e7d2aa376"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.794315 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.794796 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp6sp\" (UniqueName: \"kubernetes.io/projected/e2166478-a6f0-421b-9ea7-17f60ee9605b-kube-api-access-xp6sp\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.794856 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.794908 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02360792-0295-42be-a94c-de0e7d2aa376-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.794992 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.795860 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2166478-a6f0-421b-9ea7-17f60ee9605b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.796043 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.853651 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.862431 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2166478-a6f0-421b-9ea7-17f60ee9605b" (UID: "e2166478-a6f0-421b-9ea7-17f60ee9605b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.901721 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.901754 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.904313 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-config-data" (OuterVolumeSpecName: "config-data") pod "e2166478-a6f0-421b-9ea7-17f60ee9605b" (UID: "e2166478-a6f0-421b-9ea7-17f60ee9605b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:10 crc kubenswrapper[4788]: I1010 15:09:10.913067 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e2166478-a6f0-421b-9ea7-17f60ee9605b" (UID: "e2166478-a6f0-421b-9ea7-17f60ee9605b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.003897 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.004258 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2166478-a6f0-421b-9ea7-17f60ee9605b-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.083316 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.107404 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.113461 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-pz5rc], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone56b0-account-delete-hfwph" podUID="3aa6d9f2-ee07-4cff-9cdc-def5e94ede33" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.121236 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.122435 4788 scope.go:117] "RemoveContainer" containerID="507ce22908f8278601f3c702001bc28ce6a57e91e11224453c2ba8d3538b36de" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.148291 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.160495 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican59ec-account-delete-nn85b" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.174719 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.187731 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.203951 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.215150 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-httpd-run\") pod \"df85b9e1-2830-40c3-bbea-555bcddf5545\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.215242 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67000f4b-b955-44d7-ad37-3d552d8efd88-logs\") pod \"67000f4b-b955-44d7-ad37-3d552d8efd88\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.215273 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-public-tls-certs\") pod \"67000f4b-b955-44d7-ad37-3d552d8efd88\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.215301 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-internal-tls-certs\") pod \"df85b9e1-2830-40c3-bbea-555bcddf5545\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.216459 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67000f4b-b955-44d7-ad37-3d552d8efd88-logs" (OuterVolumeSpecName: "logs") pod "67000f4b-b955-44d7-ad37-3d552d8efd88" (UID: "67000f4b-b955-44d7-ad37-3d552d8efd88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.217957 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "df85b9e1-2830-40c3-bbea-555bcddf5545" (UID: "df85b9e1-2830-40c3-bbea-555bcddf5545"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.273940 4788 scope.go:117] "RemoveContainer" containerID="62068765dbbe9a789befff7b25adc8fe43b248f71c27fd2edbc0898452134000" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.276764 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-scripts\") pod \"67000f4b-b955-44d7-ad37-3d552d8efd88\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.276812 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-scripts\") pod \"df85b9e1-2830-40c3-bbea-555bcddf5545\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.276838 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwbj2\" (UniqueName: \"kubernetes.io/projected/67000f4b-b955-44d7-ad37-3d552d8efd88-kube-api-access-vwbj2\") pod \"67000f4b-b955-44d7-ad37-3d552d8efd88\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.276867 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-internal-tls-certs\") pod \"67000f4b-b955-44d7-ad37-3d552d8efd88\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.276894 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-combined-ca-bundle\") pod \"67000f4b-b955-44d7-ad37-3d552d8efd88\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.276925 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw8hf\" (UniqueName: \"kubernetes.io/projected/df85b9e1-2830-40c3-bbea-555bcddf5545-kube-api-access-lw8hf\") pod \"df85b9e1-2830-40c3-bbea-555bcddf5545\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.276959 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-combined-ca-bundle\") pod \"df85b9e1-2830-40c3-bbea-555bcddf5545\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.277053 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-config-data\") pod \"4fad8d4a-3435-4647-b10d-51951c64e49a\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.277088 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-config-data\") pod \"67000f4b-b955-44d7-ad37-3d552d8efd88\" (UID: \"67000f4b-b955-44d7-ad37-3d552d8efd88\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.277110 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-logs\") pod \"df85b9e1-2830-40c3-bbea-555bcddf5545\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.277131 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d2fx\" (UniqueName: \"kubernetes.io/projected/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-kube-api-access-2d2fx\") pod \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.277232 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-config-data\") pod \"df85b9e1-2830-40c3-bbea-555bcddf5545\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.277258 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.277289 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"df85b9e1-2830-40c3-bbea-555bcddf5545\" (UID: \"df85b9e1-2830-40c3-bbea-555bcddf5545\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.284700 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-scripts" (OuterVolumeSpecName: "scripts") pod "67000f4b-b955-44d7-ad37-3d552d8efd88" (UID: "67000f4b-b955-44d7-ad37-3d552d8efd88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.287579 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-scripts" (OuterVolumeSpecName: "scripts") pod "df85b9e1-2830-40c3-bbea-555bcddf5545" (UID: "df85b9e1-2830-40c3-bbea-555bcddf5545"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.288005 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-logs" (OuterVolumeSpecName: "logs") pod "df85b9e1-2830-40c3-bbea-555bcddf5545" (UID: "df85b9e1-2830-40c3-bbea-555bcddf5545"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.288774 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.289247 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.289456 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data podName:dcf4b178-75ee-47fb-81e4-e6b277eb1790 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:19.289372193 +0000 UTC m=+1461.739087741 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data") pod "rabbitmq-server-0" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790") : configmap "rabbitmq-config-data" not found Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.295094 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-kube-api-access-2d2fx" (OuterVolumeSpecName: "kube-api-access-2d2fx") pod "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" (UID: "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8"). InnerVolumeSpecName "kube-api-access-2d2fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.296999 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67000f4b-b955-44d7-ad37-3d552d8efd88-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.300635 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "df85b9e1-2830-40c3-bbea-555bcddf5545" (UID: "df85b9e1-2830-40c3-bbea-555bcddf5545"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.315043 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67000f4b-b955-44d7-ad37-3d552d8efd88-kube-api-access-vwbj2" (OuterVolumeSpecName: "kube-api-access-vwbj2") pod "67000f4b-b955-44d7-ad37-3d552d8efd88" (UID: "67000f4b-b955-44d7-ad37-3d552d8efd88"). InnerVolumeSpecName "kube-api-access-vwbj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.317030 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.320893 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df85b9e1-2830-40c3-bbea-555bcddf5545-kube-api-access-lw8hf" (OuterVolumeSpecName: "kube-api-access-lw8hf") pod "df85b9e1-2830-40c3-bbea-555bcddf5545" (UID: "df85b9e1-2830-40c3-bbea-555bcddf5545"). InnerVolumeSpecName "kube-api-access-lw8hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.329226 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.345267 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.351236 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-config-data" (OuterVolumeSpecName: "config-data") pod "4fad8d4a-3435-4647-b10d-51951c64e49a" (UID: "4fad8d4a-3435-4647-b10d-51951c64e49a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.353713 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-5b795d78cf-6sr7n"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.373202 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-5b795d78cf-6sr7n"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.380961 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.382622 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "df85b9e1-2830-40c3-bbea-555bcddf5545" (UID: "df85b9e1-2830-40c3-bbea-555bcddf5545"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.391411 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.391579 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement4ba5-account-delete-ljq8m"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.399330 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc9dc\" (UniqueName: \"kubernetes.io/projected/4aeeba9b-9894-4293-a673-7e88c6bc80f5-kube-api-access-jc9dc\") pod \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.400872 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6kwj\" (UniqueName: \"kubernetes.io/projected/5bfee301-605c-4fb8-992e-8911bfc7206f-kube-api-access-t6kwj\") pod \"5bfee301-605c-4fb8-992e-8911bfc7206f\" (UID: \"5bfee301-605c-4fb8-992e-8911bfc7206f\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.401611 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-internal-tls-certs\") pod \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.401724 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-config-data\") pod \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.401810 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2qw6\" (UniqueName: \"kubernetes.io/projected/7b604164-1660-4f3a-929b-cf6f59b66823-kube-api-access-n2qw6\") pod \"7b604164-1660-4f3a-929b-cf6f59b66823\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.401893 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-public-tls-certs\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.401968 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-combined-ca-bundle\") pod \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.407763 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-combined-ca-bundle\") pod \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.407936 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-scripts\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.410851 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-logs\") pod \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.410985 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-internal-tls-certs\") pod \"7b604164-1660-4f3a-929b-cf6f59b66823\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.411072 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-combined-ca-bundle\") pod \"7b604164-1660-4f3a-929b-cf6f59b66823\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.411190 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-scripts\") pod \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.414308 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-public-tls-certs\") pod \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.414490 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6sj2\" (UniqueName: \"kubernetes.io/projected/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-kube-api-access-v6sj2\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.414622 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-logs\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.414770 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88574\" (UniqueName: \"kubernetes.io/projected/4fad8d4a-3435-4647-b10d-51951c64e49a-kube-api-access-88574\") pod \"4fad8d4a-3435-4647-b10d-51951c64e49a\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.414884 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-logs\") pod \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.414976 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-nova-metadata-tls-certs\") pod \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\" (UID: \"f2d4fc78-0ea7-49fe-9da3-caed0438e9f8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.415074 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-public-tls-certs\") pod \"7b604164-1660-4f3a-929b-cf6f59b66823\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.416174 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-config-data\") pod \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.416295 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-combined-ca-bundle\") pod \"4fad8d4a-3435-4647-b10d-51951c64e49a\" (UID: \"4fad8d4a-3435-4647-b10d-51951c64e49a\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.416414 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-combined-ca-bundle\") pod \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.416546 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data\") pod \"7b604164-1660-4f3a-929b-cf6f59b66823\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.416658 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcxnq\" (UniqueName: \"kubernetes.io/projected/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-kube-api-access-bcxnq\") pod \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\" (UID: \"4b7f2c05-93d2-4139-95ea-c36bffcecdb8\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.416799 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-internal-tls-certs\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.416906 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-etc-machine-id\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.417076 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-combined-ca-bundle\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.419067 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-config-data\") pod \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.419228 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-sg-core-conf-yaml\") pod \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.419331 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-run-httpd\") pod \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.419434 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data-custom\") pod \"7b604164-1660-4f3a-929b-cf6f59b66823\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.419537 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b604164-1660-4f3a-929b-cf6f59b66823-logs\") pod \"7b604164-1660-4f3a-929b-cf6f59b66823\" (UID: \"7b604164-1660-4f3a-929b-cf6f59b66823\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.419645 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data-custom\") pod \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\" (UID: \"ab6c469a-3a29-4fdb-b0ce-5671db6645d2\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.419752 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-ceilometer-tls-certs\") pod \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.420519 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz5rc\" (UniqueName: \"kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc\") pod \"keystone56b0-account-delete-hfwph\" (UID: \"3aa6d9f2-ee07-4cff-9cdc-def5e94ede33\") " pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.420746 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.420819 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.420903 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.420982 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwbj2\" (UniqueName: \"kubernetes.io/projected/67000f4b-b955-44d7-ad37-3d552d8efd88-kube-api-access-vwbj2\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.421061 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw8hf\" (UniqueName: \"kubernetes.io/projected/df85b9e1-2830-40c3-bbea-555bcddf5545-kube-api-access-lw8hf\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.421134 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.421255 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df85b9e1-2830-40c3-bbea-555bcddf5545-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.421464 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d2fx\" (UniqueName: \"kubernetes.io/projected/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-kube-api-access-2d2fx\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.421569 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.431644 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4aeeba9b-9894-4293-a673-7e88c6bc80f5" (UID: "4aeeba9b-9894-4293-a673-7e88c6bc80f5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.434801 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b604164-1660-4f3a-929b-cf6f59b66823-logs" (OuterVolumeSpecName: "logs") pod "7b604164-1660-4f3a-929b-cf6f59b66823" (UID: "7b604164-1660-4f3a-929b-cf6f59b66823"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.420909 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement4ba5-account-delete-ljq8m"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.435593 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-87b66898f-smc95"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.435618 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-87b66898f-smc95"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.435908 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-logs" (OuterVolumeSpecName: "logs") pod "4b7f2c05-93d2-4139-95ea-c36bffcecdb8" (UID: "4b7f2c05-93d2-4139-95ea-c36bffcecdb8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.435959 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.436446 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-logs" (OuterVolumeSpecName: "logs") pod "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" (UID: "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.438978 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aeeba9b-9894-4293-a673-7e88c6bc80f5-kube-api-access-jc9dc" (OuterVolumeSpecName: "kube-api-access-jc9dc") pod "4aeeba9b-9894-4293-a673-7e88c6bc80f5" (UID: "4aeeba9b-9894-4293-a673-7e88c6bc80f5"). InnerVolumeSpecName "kube-api-access-jc9dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.440082 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-logs" (OuterVolumeSpecName: "logs") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.441944 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.445415 4788 projected.go:194] Error preparing data for projected volume kube-api-access-pz5rc for pod openstack/keystone56b0-account-delete-hfwph: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.446573 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc podName:3aa6d9f2-ee07-4cff-9cdc-def5e94ede33 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:13.44655252 +0000 UTC m=+1455.896268068 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-pz5rc" (UniqueName: "kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc") pod "keystone56b0-account-delete-hfwph" (UID: "3aa6d9f2-ee07-4cff-9cdc-def5e94ede33") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.446932 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-scripts" (OuterVolumeSpecName: "scripts") pod "4aeeba9b-9894-4293-a673-7e88c6bc80f5" (UID: "4aeeba9b-9894-4293-a673-7e88c6bc80f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.451998 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7b604164-1660-4f3a-929b-cf6f59b66823" (UID: "7b604164-1660-4f3a-929b-cf6f59b66823"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.453222 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.455263 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bfee301-605c-4fb8-992e-8911bfc7206f-kube-api-access-t6kwj" (OuterVolumeSpecName: "kube-api-access-t6kwj") pod "5bfee301-605c-4fb8-992e-8911bfc7206f" (UID: "5bfee301-605c-4fb8-992e-8911bfc7206f"). InnerVolumeSpecName "kube-api-access-t6kwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.466424 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderb986-account-delete-tvgzp"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.469992 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-kube-api-access-bcxnq" (OuterVolumeSpecName: "kube-api-access-bcxnq") pod "4b7f2c05-93d2-4139-95ea-c36bffcecdb8" (UID: "4b7f2c05-93d2-4139-95ea-c36bffcecdb8"). InnerVolumeSpecName "kube-api-access-bcxnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.470061 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.470099 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fad8d4a-3435-4647-b10d-51951c64e49a-kube-api-access-88574" (OuterVolumeSpecName: "kube-api-access-88574") pod "4fad8d4a-3435-4647-b10d-51951c64e49a" (UID: "4fad8d4a-3435-4647-b10d-51951c64e49a"). InnerVolumeSpecName "kube-api-access-88574". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.470227 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-kube-api-access-v6sj2" (OuterVolumeSpecName: "kube-api-access-v6sj2") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "kube-api-access-v6sj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.472356 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-scripts" (OuterVolumeSpecName: "scripts") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.472495 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b604164-1660-4f3a-929b-cf6f59b66823-kube-api-access-n2qw6" (OuterVolumeSpecName: "kube-api-access-n2qw6") pod "7b604164-1660-4f3a-929b-cf6f59b66823" (UID: "7b604164-1660-4f3a-929b-cf6f59b66823"). InnerVolumeSpecName "kube-api-access-n2qw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.515464 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.524275 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data" (OuterVolumeSpecName: "config-data") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.527221 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-log-httpd\") pod \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\" (UID: \"4aeeba9b-9894-4293-a673-7e88c6bc80f5\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528518 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcxnq\" (UniqueName: \"kubernetes.io/projected/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-kube-api-access-bcxnq\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528536 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528547 4788 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528559 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528570 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528582 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b604164-1660-4f3a-929b-cf6f59b66823-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528593 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528602 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc9dc\" (UniqueName: \"kubernetes.io/projected/4aeeba9b-9894-4293-a673-7e88c6bc80f5-kube-api-access-jc9dc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528617 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6kwj\" (UniqueName: \"kubernetes.io/projected/5bfee301-605c-4fb8-992e-8911bfc7206f-kube-api-access-t6kwj\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528628 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2qw6\" (UniqueName: \"kubernetes.io/projected/7b604164-1660-4f3a-929b-cf6f59b66823-kube-api-access-n2qw6\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528637 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528645 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528653 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528662 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6sj2\" (UniqueName: \"kubernetes.io/projected/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-kube-api-access-v6sj2\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528670 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528681 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88574\" (UniqueName: \"kubernetes.io/projected/4fad8d4a-3435-4647-b10d-51951c64e49a-kube-api-access-88574\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.528690 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-logs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.530183 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4aeeba9b-9894-4293-a673-7e88c6bc80f5" (UID: "4aeeba9b-9894-4293-a673-7e88c6bc80f5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.535930 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinderb986-account-delete-tvgzp"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.541785 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.543353 4788 scope.go:117] "RemoveContainer" containerID="87673bff375f480e6dff20a3a6b396e073584e22ecfa29abe8788960477c7c73" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.592981 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.613941 4788 scope.go:117] "RemoveContainer" containerID="fd696a35a1e5c7857218ceaedea76b07e06bb96d506d78970a1c5927133e9d90" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.618667 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.629735 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crw2j\" (UniqueName: \"kubernetes.io/projected/cf293818-85fb-46dc-8cee-ba8eca827bb7-kube-api-access-crw2j\") pod \"cf293818-85fb-46dc-8cee-ba8eca827bb7\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.630570 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-combined-ca-bundle\") pod \"cf293818-85fb-46dc-8cee-ba8eca827bb7\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.630672 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-config-data\") pod \"cf293818-85fb-46dc-8cee-ba8eca827bb7\" (UID: \"cf293818-85fb-46dc-8cee-ba8eca827bb7\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.631099 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4aeeba9b-9894-4293-a673-7e88c6bc80f5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.631113 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.648814 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.669044 4788 scope.go:117] "RemoveContainer" containerID="06b8e0b90057fe848ada9d1d4fb30735ed3b1c3365172a8e3399572d5e67a2a7" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.670559 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf293818-85fb-46dc-8cee-ba8eca827bb7-kube-api-access-crw2j" (OuterVolumeSpecName: "kube-api-access-crw2j") pod "cf293818-85fb-46dc-8cee-ba8eca827bb7" (UID: "cf293818-85fb-46dc-8cee-ba8eca827bb7"). InnerVolumeSpecName "kube-api-access-crw2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.696010 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-657c889cbb-d8d5h"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.704084 4788 scope.go:117] "RemoveContainer" containerID="ec7fa8d6b4673ceda9dd84c9e935b5a50e9b9b7c096828ed3a77bd95edb51e09" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.714194 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" (UID: "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.715915 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf293818-85fb-46dc-8cee-ba8eca827bb7" (UID: "cf293818-85fb-46dc-8cee-ba8eca827bb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.729348 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-657c889cbb-d8d5h"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.734089 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-config-data\") pod \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.737058 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kolla-config\") pod \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.737461 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-memcached-tls-certs\") pod \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.740891 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgk4f\" (UniqueName: \"kubernetes.io/projected/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kube-api-access-zgk4f\") pod \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.743453 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-combined-ca-bundle\") pod \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\" (UID: \"4075bf4d-01ed-4d7c-9fb8-576dd2866821\") " Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.744885 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crw2j\" (UniqueName: \"kubernetes.io/projected/cf293818-85fb-46dc-8cee-ba8eca827bb7-kube-api-access-crw2j\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.744915 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.744930 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.737835 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-config-data" (OuterVolumeSpecName: "config-data") pod "4075bf4d-01ed-4d7c-9fb8-576dd2866821" (UID: "4075bf4d-01ed-4d7c-9fb8-576dd2866821"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.738156 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "4075bf4d-01ed-4d7c-9fb8-576dd2866821" (UID: "4075bf4d-01ed-4d7c-9fb8-576dd2866821"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.745005 4788 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.745131 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data podName:a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c nodeName:}" failed. No retries permitted until 2025-10-10 15:09:19.745112555 +0000 UTC m=+1462.194828103 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data") pod "rabbitmq-cell1-server-0" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c") : configmap "rabbitmq-cell1-config-data" not found Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.760132 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.763804 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.794675 4788 scope.go:117] "RemoveContainer" containerID="737b4bebc0138dcddcb915236eb4ef30c70a15399b522e7c7159d53fa58d0ee4" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.806285 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kube-api-access-zgk4f" (OuterVolumeSpecName: "kube-api-access-zgk4f") pod "4075bf4d-01ed-4d7c-9fb8-576dd2866821" (UID: "4075bf4d-01ed-4d7c-9fb8-576dd2866821"). InnerVolumeSpecName "kube-api-access-zgk4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.813634 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754 is running failed: container process not found" containerID="ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.815369 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754 is running failed: container process not found" containerID="ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.823197 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754 is running failed: container process not found" containerID="ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.823269 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="ovn-northd" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.835867 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b7f2c05-93d2-4139-95ea-c36bffcecdb8" (UID: "4b7f2c05-93d2-4139-95ea-c36bffcecdb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.848830 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0bb0b-account-delete-lcsdj" event={"ID":"92ccaefc-72b3-4e2f-90c0-394240ed2ab2","Type":"ContainerStarted","Data":"39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb"} Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.849367 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell0bb0b-account-delete-lcsdj" podUID="92ccaefc-72b3-4e2f-90c0-394240ed2ab2" containerName="mariadb-account-delete" containerID="cri-o://39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb" gracePeriod=30 Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.853535 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronab26-account-delete-g2dnr" event={"ID":"88f8c7ed-d478-4bae-ae23-4bffcb16abf6","Type":"ContainerStarted","Data":"d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259"} Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.854027 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutronab26-account-delete-g2dnr" podUID="88f8c7ed-d478-4bae-ae23-4bffcb16abf6" containerName="mariadb-account-delete" containerID="cri-o://d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259" gracePeriod=30 Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.863956 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.863995 4788 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.864005 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgk4f\" (UniqueName: \"kubernetes.io/projected/4075bf4d-01ed-4d7c-9fb8-576dd2866821-kube-api-access-zgk4f\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.864015 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.875362 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.879539 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell0bb0b-account-delete-lcsdj" podStartSLOduration=8.879515186999999 podStartE2EDuration="8.879515187s" podCreationTimestamp="2025-10-10 15:09:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:09:11.867827522 +0000 UTC m=+1454.317543080" watchObservedRunningTime="2025-10-10 15:09:11.879515187 +0000 UTC m=+1454.329230735" Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.883218 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.884527 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutronab26-account-delete-g2dnr" podStartSLOduration=8.884507271 podStartE2EDuration="8.884507271s" podCreationTimestamp="2025-10-10 15:09:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:09:11.883921885 +0000 UTC m=+1454.333637423" watchObservedRunningTime="2025-10-10 15:09:11.884507271 +0000 UTC m=+1454.334222819" Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.885289 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 10 15:09:11 crc kubenswrapper[4788]: E1010 15:09:11.885861 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerName="galera" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.886293 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi96b1-account-delete-pm4dv" event={"ID":"0d375458-0dc4-45a2-abf9-c7376510d2f6","Type":"ContainerStarted","Data":"5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76"} Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.886447 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapi96b1-account-delete-pm4dv" podUID="0d375458-0dc4-45a2-abf9-c7376510d2f6" containerName="mariadb-account-delete" containerID="cri-o://5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76" gracePeriod=30 Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.894469 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a9adb465-b660-4d2c-bae8-f0b8024761b8/ovn-northd/0.log" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.894515 4788 generic.go:334] "Generic (PLEG): container finished" podID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerID="ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" exitCode=139 Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.894571 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a9adb465-b660-4d2c-bae8-f0b8024761b8","Type":"ContainerDied","Data":"ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754"} Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.909936 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cf293818-85fb-46dc-8cee-ba8eca827bb7","Type":"ContainerDied","Data":"6be74aa1cff12947235f803c8b52f306fe46a6930bef9bbe5e4b0dbce0cd3e05"} Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.910031 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.911512 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapi96b1-account-delete-pm4dv" podStartSLOduration=8.911490758 podStartE2EDuration="8.911490758s" podCreationTimestamp="2025-10-10 15:09:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:09:11.907235954 +0000 UTC m=+1454.356951502" watchObservedRunningTime="2025-10-10 15:09:11.911490758 +0000 UTC m=+1454.361206306" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.917189 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-config-data" (OuterVolumeSpecName: "config-data") pod "4b7f2c05-93d2-4139-95ea-c36bffcecdb8" (UID: "4b7f2c05-93d2-4139-95ea-c36bffcecdb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.927304 4788 generic.go:334] "Generic (PLEG): container finished" podID="4075bf4d-01ed-4d7c-9fb8-576dd2866821" containerID="2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9" exitCode=0 Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.927481 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.927544 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4075bf4d-01ed-4d7c-9fb8-576dd2866821","Type":"ContainerDied","Data":"2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9"} Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.928583 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4075bf4d-01ed-4d7c-9fb8-576dd2866821","Type":"ContainerDied","Data":"1fc4538efddc96abbaea1530461b273bd133f7132bfb713b71d2632c7c770583"} Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.937101 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b604164-1660-4f3a-929b-cf6f59b66823" (UID: "7b604164-1660-4f3a-929b-cf6f59b66823"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.939720 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance9145-account-delete-kblv8" podUID="42d4825c-6e5c-4483-ae09-657903e6acb0" containerName="mariadb-account-delete" containerID="cri-o://5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b" gracePeriod=30 Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.939945 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance9145-account-delete-kblv8" event={"ID":"42d4825c-6e5c-4483-ae09-657903e6acb0","Type":"ContainerStarted","Data":"5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b"} Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.957091 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.957271 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-config-data" (OuterVolumeSpecName: "config-data") pod "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" (UID: "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.960804 4788 scope.go:117] "RemoveContainer" containerID="82415a1ab836383b16f8eb7bf4bf4292b585bf668709c236d3369c260e68828a" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.966076 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.966117 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.966130 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.966159 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.969353 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.969672 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5d565c7b56-z7s74" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.969824 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.969941 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican59ec-account-delete-nn85b" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.969962 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.969973 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.970010 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.970048 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.970106 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.970498 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bc8db69bb-s2ddx" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.971051 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance9145-account-delete-kblv8" podStartSLOduration=8.971031512 podStartE2EDuration="8.971031512s" podCreationTimestamp="2025-10-10 15:09:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:09:11.96239221 +0000 UTC m=+1454.412107758" watchObservedRunningTime="2025-10-10 15:09:11.971031512 +0000 UTC m=+1454.420747060" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.990167 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4075bf4d-01ed-4d7c-9fb8-576dd2866821" (UID: "4075bf4d-01ed-4d7c-9fb8-576dd2866821"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:11 crc kubenswrapper[4788]: I1010 15:09:11.990266 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-config-data" (OuterVolumeSpecName: "config-data") pod "67000f4b-b955-44d7-ad37-3d552d8efd88" (UID: "67000f4b-b955-44d7-ad37-3d552d8efd88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.000458 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67000f4b-b955-44d7-ad37-3d552d8efd88" (UID: "67000f4b-b955-44d7-ad37-3d552d8efd88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.055615 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-config-data" (OuterVolumeSpecName: "config-data") pod "cf293818-85fb-46dc-8cee-ba8eca827bb7" (UID: "cf293818-85fb-46dc-8cee-ba8eca827bb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.069353 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-config-data" (OuterVolumeSpecName: "config-data") pod "df85b9e1-2830-40c3-bbea-555bcddf5545" (UID: "df85b9e1-2830-40c3-bbea-555bcddf5545"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.076993 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.077270 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.077360 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.077509 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.077567 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf293818-85fb-46dc-8cee-ba8eca827bb7-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.100337 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df85b9e1-2830-40c3-bbea-555bcddf5545" (UID: "df85b9e1-2830-40c3-bbea-555bcddf5545"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.107515 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4aeeba9b-9894-4293-a673-7e88c6bc80f5" (UID: "4aeeba9b-9894-4293-a673-7e88c6bc80f5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.136768 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4fad8d4a-3435-4647-b10d-51951c64e49a" (UID: "4fad8d4a-3435-4647-b10d-51951c64e49a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.137568 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "67000f4b-b955-44d7-ad37-3d552d8efd88" (UID: "67000f4b-b955-44d7-ad37-3d552d8efd88"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.144289 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4b7f2c05-93d2-4139-95ea-c36bffcecdb8" (UID: "4b7f2c05-93d2-4139-95ea-c36bffcecdb8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.149909 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4b7f2c05-93d2-4139-95ea-c36bffcecdb8" (UID: "4b7f2c05-93d2-4139-95ea-c36bffcecdb8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.149908 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7b604164-1660-4f3a-929b-cf6f59b66823" (UID: "7b604164-1660-4f3a-929b-cf6f59b66823"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.171264 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data" (OuterVolumeSpecName: "config-data") pod "7b604164-1660-4f3a-929b-cf6f59b66823" (UID: "7b604164-1660-4f3a-929b-cf6f59b66823"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.171678 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-config-data" (OuterVolumeSpecName: "config-data") pod "4aeeba9b-9894-4293-a673-7e88c6bc80f5" (UID: "4aeeba9b-9894-4293-a673-7e88c6bc80f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181491 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181536 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fad8d4a-3435-4647-b10d-51951c64e49a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181546 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181556 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181565 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181575 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181585 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181594 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b7f2c05-93d2-4139-95ea-c36bffcecdb8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.181604 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df85b9e1-2830-40c3-bbea-555bcddf5545-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.183290 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.186097 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" (UID: "f2d4fc78-0ea7-49fe-9da3-caed0438e9f8"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.191715 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ab6c469a-3a29-4fdb-b0ce-5671db6645d2" (UID: "ab6c469a-3a29-4fdb-b0ce-5671db6645d2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.195253 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4aeeba9b-9894-4293-a673-7e88c6bc80f5" (UID: "4aeeba9b-9894-4293-a673-7e88c6bc80f5"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.240121 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "4075bf4d-01ed-4d7c-9fb8-576dd2866821" (UID: "4075bf4d-01ed-4d7c-9fb8-576dd2866821"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.243157 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7b604164-1660-4f3a-929b-cf6f59b66823" (UID: "7b604164-1660-4f3a-929b-cf6f59b66823"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.256131 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "67000f4b-b955-44d7-ad37-3d552d8efd88" (UID: "67000f4b-b955-44d7-ad37-3d552d8efd88"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.260826 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aeeba9b-9894-4293-a673-7e88c6bc80f5" (UID: "4aeeba9b-9894-4293-a673-7e88c6bc80f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.289665 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67000f4b-b955-44d7-ad37-3d552d8efd88-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.289719 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7b604164-1660-4f3a-929b-cf6f59b66823-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.289735 4788 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.289757 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.289772 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.289788 4788 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4aeeba9b-9894-4293-a673-7e88c6bc80f5-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.290003 4788 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4075bf4d-01ed-4d7c-9fb8-576dd2866821-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.290020 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6c469a-3a29-4fdb-b0ce-5671db6645d2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.293492 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02360792-0295-42be-a94c-de0e7d2aa376" path="/var/lib/kubelet/pods/02360792-0295-42be-a94c-de0e7d2aa376/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.294453 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="024ba580-feec-4b2a-a836-942126b022ae" path="/var/lib/kubelet/pods/024ba580-feec-4b2a-a836-942126b022ae/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.295294 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="085a2986-2c31-42f5-95c2-55f2414e5dfc" path="/var/lib/kubelet/pods/085a2986-2c31-42f5-95c2-55f2414e5dfc/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.297280 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b927859-0238-46f5-90aa-4afa094c434e" path="/var/lib/kubelet/pods/4b927859-0238-46f5-90aa-4afa094c434e/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.298246 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="505a880b-6b48-498d-aad2-b30f20d089b6" path="/var/lib/kubelet/pods/505a880b-6b48-498d-aad2-b30f20d089b6/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.299051 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f13a166-2b9b-4679-877b-f4596b2d45ed" path="/var/lib/kubelet/pods/6f13a166-2b9b-4679-877b-f4596b2d45ed/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.302930 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7810b9ee-1996-4535-bcac-5ffa8781295b" path="/var/lib/kubelet/pods/7810b9ee-1996-4535-bcac-5ffa8781295b/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.312015 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8951c1b9-28fd-449d-ade9-a7b7bfc3af6e" path="/var/lib/kubelet/pods/8951c1b9-28fd-449d-ade9-a7b7bfc3af6e/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.312888 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd8f08dc-5aaa-4429-8f8d-190a36d5f38f" path="/var/lib/kubelet/pods/cd8f08dc-5aaa-4429-8f8d-190a36d5f38f/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.313776 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" path="/var/lib/kubelet/pods/e2166478-a6f0-421b-9ea7-17f60ee9605b/volumes" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.457884 4788 scope.go:117] "RemoveContainer" containerID="705a07a6c50b14bb6069ed9b5c58612e9e5861f6620fb1c72d2da2cfd85e0e7b" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.473345 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.493295 4788 scope.go:117] "RemoveContainer" containerID="3c499468e5604cdfd26560a7f04ecbdd2ff30447627e047d1fc5e0e0d7e4cd3b" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.547243 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a9adb465-b660-4d2c-bae8-f0b8024761b8/ovn-northd/0.log" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.547395 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.553978 4788 scope.go:117] "RemoveContainer" containerID="22d302e6d9d9eed7daebd5122f717fd594b1258085c469fab8d22ea7ffda2b6e" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.569544 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594541 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594600 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-scripts\") pod \"a9adb465-b660-4d2c-bae8-f0b8024761b8\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594634 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-erlang-cookie\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594681 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dcf4b178-75ee-47fb-81e4-e6b277eb1790-pod-info\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594721 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-rundir\") pod \"a9adb465-b660-4d2c-bae8-f0b8024761b8\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594777 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-config\") pod \"a9adb465-b660-4d2c-bae8-f0b8024761b8\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594840 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-northd-tls-certs\") pod \"a9adb465-b660-4d2c-bae8-f0b8024761b8\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594897 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-plugins-conf\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.594961 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-server-conf\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595058 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55n9j\" (UniqueName: \"kubernetes.io/projected/a9adb465-b660-4d2c-bae8-f0b8024761b8-kube-api-access-55n9j\") pod \"a9adb465-b660-4d2c-bae8-f0b8024761b8\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595085 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-plugins\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595102 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x97t\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-kube-api-access-2x97t\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595132 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-tls\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595178 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dcf4b178-75ee-47fb-81e4-e6b277eb1790-erlang-cookie-secret\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595521 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-confd\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595635 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-metrics-certs-tls-certs\") pod \"a9adb465-b660-4d2c-bae8-f0b8024761b8\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595709 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data\") pod \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\" (UID: \"dcf4b178-75ee-47fb-81e4-e6b277eb1790\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.595855 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-combined-ca-bundle\") pod \"a9adb465-b660-4d2c-bae8-f0b8024761b8\" (UID: \"a9adb465-b660-4d2c-bae8-f0b8024761b8\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.598272 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.601342 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a9adb465-b660-4d2c-bae8-f0b8024761b8" (UID: "a9adb465-b660-4d2c-bae8-f0b8024761b8"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.609504 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-config" (OuterVolumeSpecName: "config") pod "a9adb465-b660-4d2c-bae8-f0b8024761b8" (UID: "a9adb465-b660-4d2c-bae8-f0b8024761b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.610383 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.612440 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.612506 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.611905 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-scripts" (OuterVolumeSpecName: "scripts") pod "a9adb465-b660-4d2c-bae8-f0b8024761b8" (UID: "a9adb465-b660-4d2c-bae8-f0b8024761b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.616521 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.624033 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9adb465-b660-4d2c-bae8-f0b8024761b8-kube-api-access-55n9j" (OuterVolumeSpecName: "kube-api-access-55n9j") pod "a9adb465-b660-4d2c-bae8-f0b8024761b8" (UID: "a9adb465-b660-4d2c-bae8-f0b8024761b8"). InnerVolumeSpecName "kube-api-access-55n9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.643640 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.644083 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.647179 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/dcf4b178-75ee-47fb-81e4-e6b277eb1790-pod-info" (OuterVolumeSpecName: "pod-info") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.652971 4788 scope.go:117] "RemoveContainer" containerID="1030458b8b3ff8477d57735857a468ea3a2e6bd973b1045fcf998e6c108d3b3f" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.660395 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-bc8db69bb-s2ddx"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.670646 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-bc8db69bb-s2ddx"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.684392 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-kube-api-access-2x97t" (OuterVolumeSpecName: "kube-api-access-2x97t") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "kube-api-access-2x97t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.690818 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.713632 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717567 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717608 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717625 4788 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dcf4b178-75ee-47fb-81e4-e6b277eb1790-pod-info\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717639 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717666 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9adb465-b660-4d2c-bae8-f0b8024761b8-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717685 4788 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717698 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55n9j\" (UniqueName: \"kubernetes.io/projected/a9adb465-b660-4d2c-bae8-f0b8024761b8-kube-api-access-55n9j\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717712 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.717726 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x97t\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-kube-api-access-2x97t\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.718303 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.723725 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcf4b178-75ee-47fb-81e4-e6b277eb1790-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.735417 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.759029 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5d565c7b56-z7s74"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.759475 4788 scope.go:117] "RemoveContainer" containerID="d0bc3210873389e2972e8761b9e10f4be661565d640a62e62ffc9f730bc70876" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.765177 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5d565c7b56-z7s74"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.806229 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.824947 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-combined-ca-bundle\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825004 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825039 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-generated\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825062 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-default\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825111 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-secrets\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825129 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-operator-scripts\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825199 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-galera-tls-certs\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825224 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz4jq\" (UniqueName: \"kubernetes.io/projected/4c25a6d1-ee44-438c-a16d-e6952cfda420-kube-api-access-wz4jq\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825252 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-kolla-config\") pod \"4c25a6d1-ee44-438c-a16d-e6952cfda420\" (UID: \"4c25a6d1-ee44-438c-a16d-e6952cfda420\") " Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825445 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.825456 4788 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dcf4b178-75ee-47fb-81e4-e6b277eb1790-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.839959 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.840517 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.845119 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.847851 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.855104 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.862674 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data" (OuterVolumeSpecName: "config-data") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.863322 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.876443 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-secrets" (OuterVolumeSpecName: "secrets") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.883202 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c25a6d1-ee44-438c-a16d-e6952cfda420-kube-api-access-wz4jq" (OuterVolumeSpecName: "kube-api-access-wz4jq") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "kube-api-access-wz4jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.889614 4788 scope.go:117] "RemoveContainer" containerID="905b0fdae791afc90732669e1fd50c89e576f7fe20ab044f3b24b621f2e22853" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.909404 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.933385 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz4jq\" (UniqueName: \"kubernetes.io/projected/4c25a6d1-ee44-438c-a16d-e6952cfda420-kube-api-access-wz4jq\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.933417 4788 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.933427 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.933436 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.933445 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.933453 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.933462 4788 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-secrets\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.933470 4788 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c25a6d1-ee44-438c-a16d-e6952cfda420-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.974346 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.974424 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:12 crc kubenswrapper[4788]: I1010 15:09:12.974457 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9adb465-b660-4d2c-bae8-f0b8024761b8" (UID: "a9adb465-b660-4d2c-bae8-f0b8024761b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.015272 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038455 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038536 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-plugins\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038601 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-pod-info\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038650 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-erlang-cookie-secret\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038717 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-confd\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038770 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-erlang-cookie\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038838 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z55k\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-kube-api-access-8z55k\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038910 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038947 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-plugins-conf\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.038993 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-tls\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.039041 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-server-conf\") pod \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\" (UID: \"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.039341 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.039366 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.039378 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.044536 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.050964 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.052889 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.052966 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.053887 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.055901 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.073860 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.075027 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-pod-info" (OuterVolumeSpecName: "pod-info") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.077094 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.085418 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.087617 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "4c25a6d1-ee44-438c-a16d-e6952cfda420" (UID: "4c25a6d1-ee44-438c-a16d-e6952cfda420"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.092613 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a9adb465-b660-4d2c-bae8-f0b8024761b8" (UID: "a9adb465-b660-4d2c-bae8-f0b8024761b8"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.092819 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-server-conf" (OuterVolumeSpecName: "server-conf") pod "dcf4b178-75ee-47fb-81e4-e6b277eb1790" (UID: "dcf4b178-75ee-47fb-81e4-e6b277eb1790"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.093092 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.093180 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c25a6d1-ee44-438c-a16d-e6952cfda420","Type":"ContainerDied","Data":"36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74"} Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.093095 4788 generic.go:334] "Generic (PLEG): container finished" podID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerID="36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74" exitCode=0 Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.093668 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c25a6d1-ee44-438c-a16d-e6952cfda420","Type":"ContainerDied","Data":"64d5acfa5ac3affc79512c1007256b05a19ec5a3e209fe3a91578ee2901b87b3"} Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.098621 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-kube-api-access-8z55k" (OuterVolumeSpecName: "kube-api-access-8z55k") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "kube-api-access-8z55k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.105684 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.106742 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_a9adb465-b660-4d2c-bae8-f0b8024761b8/ovn-northd/0.log" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.106879 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"a9adb465-b660-4d2c-bae8-f0b8024761b8","Type":"ContainerDied","Data":"e2e5e941f3144ddea505b482e85894d2ab1422c68a655bf2103d1f36b70fffd7"} Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.112611 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.112630 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.114518 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data" (OuterVolumeSpecName: "config-data") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.118493 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican59ec-account-delete-nn85b"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.122191 4788 generic.go:334] "Generic (PLEG): container finished" podID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerID="c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251" exitCode=0 Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.122436 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.124673 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dcf4b178-75ee-47fb-81e4-e6b277eb1790","Type":"ContainerDied","Data":"c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251"} Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.124741 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dcf4b178-75ee-47fb-81e4-e6b277eb1790","Type":"ContainerDied","Data":"0b44a0731057457a37fc02194611bb667c8fdefa9a770631b91b70d997c40b65"} Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140083 4788 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140122 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140134 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140160 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140169 4788 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dcf4b178-75ee-47fb-81e4-e6b277eb1790-server-conf\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140179 4788 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-pod-info\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140189 4788 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140199 4788 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c25a6d1-ee44-438c-a16d-e6952cfda420-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140208 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dcf4b178-75ee-47fb-81e4-e6b277eb1790-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140218 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140228 4788 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140242 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z55k\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-kube-api-access-8z55k\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140252 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.140275 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.159535 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican59ec-account-delete-nn85b"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.161075 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-server-conf" (OuterVolumeSpecName: "server-conf") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.164878 4788 generic.go:334] "Generic (PLEG): container finished" podID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerID="a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab" exitCode=0 Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.164966 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c","Type":"ContainerDied","Data":"a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab"} Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.165002 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c","Type":"ContainerDied","Data":"9c0961e4c5c1b380e25b4b984eda2092f6a4523489dc2ab507510ae673855d7a"} Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.165016 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.167482 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.170869 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.172832 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "a9adb465-b660-4d2c-bae8-f0b8024761b8" (UID: "a9adb465-b660-4d2c-bae8-f0b8024761b8"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.174046 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.182193 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.196721 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" (UID: "a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.196800 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.212635 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.236471 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.242622 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9adb465-b660-4d2c-bae8-f0b8024761b8-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.242671 4788 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-server-conf\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.242684 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.243402 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.250197 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.251828 4788 scope.go:117] "RemoveContainer" containerID="b5c371e53f6216ef8edfc3b5203d8b1eccf15ab2af6828b0ddf25a2c86e85ada" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.277606 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.303895 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.312240 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.312312 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.320383 4788 scope.go:117] "RemoveContainer" containerID="1ac5b4a4856f6b0a772902a9c4e92e64c3a65186c5c7b198248ff40fef09742d" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.366189 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone56b0-account-delete-hfwph"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.373891 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone56b0-account-delete-hfwph"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.378763 4788 scope.go:117] "RemoveContainer" containerID="2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.408202 4788 scope.go:117] "RemoveContainer" containerID="2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.408947 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9\": container with ID starting with 2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9 not found: ID does not exist" containerID="2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.409021 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9"} err="failed to get container status \"2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9\": rpc error: code = NotFound desc = could not find container \"2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9\": container with ID starting with 2638c2a35c300971dddb3ee8ffce57800c2a01abe54dc64f1b2876160334d0c9 not found: ID does not exist" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.409068 4788 scope.go:117] "RemoveContainer" containerID="36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.448901 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz5rc\" (UniqueName: \"kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc\") pod \"keystone56b0-account-delete-hfwph\" (UID: \"3aa6d9f2-ee07-4cff-9cdc-def5e94ede33\") " pod="openstack/keystone56b0-account-delete-hfwph" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.451526 4788 projected.go:194] Error preparing data for projected volume kube-api-access-pz5rc for pod openstack/keystone56b0-account-delete-hfwph: failed to fetch token: pod "keystone56b0-account-delete-hfwph" not found Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.451600 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc podName:3aa6d9f2-ee07-4cff-9cdc-def5e94ede33 nodeName:}" failed. No retries permitted until 2025-10-10 15:09:17.451579929 +0000 UTC m=+1459.901295477 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-pz5rc" (UniqueName: "kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc") pod "keystone56b0-account-delete-hfwph" (UID: "3aa6d9f2-ee07-4cff-9cdc-def5e94ede33") : failed to fetch token: pod "keystone56b0-account-delete-hfwph" not found Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.466108 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.472035 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.481090 4788 scope.go:117] "RemoveContainer" containerID="154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.535811 4788 scope.go:117] "RemoveContainer" containerID="36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.536502 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74\": container with ID starting with 36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74 not found: ID does not exist" containerID="36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.536555 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74"} err="failed to get container status \"36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74\": rpc error: code = NotFound desc = could not find container \"36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74\": container with ID starting with 36e26a6932648fe6fa7506ffe21a8a018f7c3ed8c70f3df2992b7795be8a7c74 not found: ID does not exist" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.536589 4788 scope.go:117] "RemoveContainer" containerID="154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.536689 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.538227 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20\": container with ID starting with 154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20 not found: ID does not exist" containerID="154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.538261 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20"} err="failed to get container status \"154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20\": rpc error: code = NotFound desc = could not find container \"154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20\": container with ID starting with 154ce91314ba4a107ebb1770316c314447a700d7675365f347dacb0c5a0ebb20 not found: ID does not exist" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.538277 4788 scope.go:117] "RemoveContainer" containerID="335000a0690e32f11a92cfd8dfc8fe055195d80c307ea823dc4da83c8237d39d" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.551874 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz5rc\" (UniqueName: \"kubernetes.io/projected/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33-kube-api-access-pz5rc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.552067 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.599312 4788 scope.go:117] "RemoveContainer" containerID="ecf64525d0e44d77a8ddd04f95d78097adc9d675abc06bf3eec12ae549ee7754" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.601012 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.641372 4788 scope.go:117] "RemoveContainer" containerID="c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.663624 4788 scope.go:117] "RemoveContainer" containerID="a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.700411 4788 scope.go:117] "RemoveContainer" containerID="c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.701520 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251\": container with ID starting with c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251 not found: ID does not exist" containerID="c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.701568 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251"} err="failed to get container status \"c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251\": rpc error: code = NotFound desc = could not find container \"c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251\": container with ID starting with c7bbbf4b471873e26d6741a7eadf4521159e5802ac8fdc6b0df37327f1e3c251 not found: ID does not exist" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.701600 4788 scope.go:117] "RemoveContainer" containerID="a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.702158 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290\": container with ID starting with a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290 not found: ID does not exist" containerID="a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.702244 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290"} err="failed to get container status \"a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290\": rpc error: code = NotFound desc = could not find container \"a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290\": container with ID starting with a6abe23cfaf5bbd9e62834961816f1c63bc83b70fa3509786ffa300d01778290 not found: ID does not exist" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.702282 4788 scope.go:117] "RemoveContainer" containerID="a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.735902 4788 scope.go:117] "RemoveContainer" containerID="54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.753976 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-fernet-keys\") pod \"970e7515-bd36-45d9-befb-a0597cc72cbb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.754070 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrsln\" (UniqueName: \"kubernetes.io/projected/970e7515-bd36-45d9-befb-a0597cc72cbb-kube-api-access-mrsln\") pod \"970e7515-bd36-45d9-befb-a0597cc72cbb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.754108 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-combined-ca-bundle\") pod \"970e7515-bd36-45d9-befb-a0597cc72cbb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.754171 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-public-tls-certs\") pod \"970e7515-bd36-45d9-befb-a0597cc72cbb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.754231 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-internal-tls-certs\") pod \"970e7515-bd36-45d9-befb-a0597cc72cbb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.754348 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-credential-keys\") pod \"970e7515-bd36-45d9-befb-a0597cc72cbb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.754416 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-config-data\") pod \"970e7515-bd36-45d9-befb-a0597cc72cbb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.754487 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-scripts\") pod \"970e7515-bd36-45d9-befb-a0597cc72cbb\" (UID: \"970e7515-bd36-45d9-befb-a0597cc72cbb\") " Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.761334 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/970e7515-bd36-45d9-befb-a0597cc72cbb-kube-api-access-mrsln" (OuterVolumeSpecName: "kube-api-access-mrsln") pod "970e7515-bd36-45d9-befb-a0597cc72cbb" (UID: "970e7515-bd36-45d9-befb-a0597cc72cbb"). InnerVolumeSpecName "kube-api-access-mrsln". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.764646 4788 scope.go:117] "RemoveContainer" containerID="a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.765328 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "970e7515-bd36-45d9-befb-a0597cc72cbb" (UID: "970e7515-bd36-45d9-befb-a0597cc72cbb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.767736 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab\": container with ID starting with a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab not found: ID does not exist" containerID="a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.767788 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab"} err="failed to get container status \"a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab\": rpc error: code = NotFound desc = could not find container \"a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab\": container with ID starting with a06f4692139feb4337f57242b7bcb72eae0acfef7c9a7f2c6d13d560bb84e6ab not found: ID does not exist" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.768270 4788 scope.go:117] "RemoveContainer" containerID="54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.768365 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-scripts" (OuterVolumeSpecName: "scripts") pod "970e7515-bd36-45d9-befb-a0597cc72cbb" (UID: "970e7515-bd36-45d9-befb-a0597cc72cbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.768434 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "970e7515-bd36-45d9-befb-a0597cc72cbb" (UID: "970e7515-bd36-45d9-befb-a0597cc72cbb"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.771796 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45\": container with ID starting with 54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45 not found: ID does not exist" containerID="54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.771834 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45"} err="failed to get container status \"54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45\": rpc error: code = NotFound desc = could not find container \"54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45\": container with ID starting with 54404af9c2847b3c9f615437b8c9f9d0fa2f394335a7e3f0fd12144825f08f45 not found: ID does not exist" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.801286 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "970e7515-bd36-45d9-befb-a0597cc72cbb" (UID: "970e7515-bd36-45d9-befb-a0597cc72cbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.817180 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "970e7515-bd36-45d9-befb-a0597cc72cbb" (UID: "970e7515-bd36-45d9-befb-a0597cc72cbb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.821846 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "970e7515-bd36-45d9-befb-a0597cc72cbb" (UID: "970e7515-bd36-45d9-befb-a0597cc72cbb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.829239 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-config-data" (OuterVolumeSpecName: "config-data") pod "970e7515-bd36-45d9-befb-a0597cc72cbb" (UID: "970e7515-bd36-45d9-befb-a0597cc72cbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.864362 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.864547 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.864572 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.864583 4788 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.864594 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.864604 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.864613 4788 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.864622 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrsln\" (UniqueName: \"kubernetes.io/projected/970e7515-bd36-45d9-befb-a0597cc72cbb-kube-api-access-mrsln\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: I1010 15:09:13.864633 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/970e7515-bd36-45d9-befb-a0597cc72cbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.864717 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.865007 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.865035 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.867174 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.869676 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.875249 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:13 crc kubenswrapper[4788]: E1010 15:09:13.875293 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.182099 4788 generic.go:334] "Generic (PLEG): container finished" podID="970e7515-bd36-45d9-befb-a0597cc72cbb" containerID="ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115" exitCode=0 Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.182191 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f68d7b48c-24qxb" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.182227 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f68d7b48c-24qxb" event={"ID":"970e7515-bd36-45d9-befb-a0597cc72cbb","Type":"ContainerDied","Data":"ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115"} Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.182307 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f68d7b48c-24qxb" event={"ID":"970e7515-bd36-45d9-befb-a0597cc72cbb","Type":"ContainerDied","Data":"10d0e0419d7726432bece8b517cfe4d81600b75496c1e64ab430542a46bbcb5b"} Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.182326 4788 scope.go:117] "RemoveContainer" containerID="ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.219331 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7f68d7b48c-24qxb"] Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.219429 4788 scope.go:117] "RemoveContainer" containerID="ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115" Oct 10 15:09:14 crc kubenswrapper[4788]: E1010 15:09:14.219837 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115\": container with ID starting with ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115 not found: ID does not exist" containerID="ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.219887 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115"} err="failed to get container status \"ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115\": rpc error: code = NotFound desc = could not find container \"ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115\": container with ID starting with ab40ad5d5e0ffb6870d3d6b4049654a88fe0936bd91b4a2496d7d7e6c5cf3115 not found: ID does not exist" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.222668 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7f68d7b48c-24qxb"] Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.243038 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aa6d9f2-ee07-4cff-9cdc-def5e94ede33" path="/var/lib/kubelet/pods/3aa6d9f2-ee07-4cff-9cdc-def5e94ede33/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.243411 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4075bf4d-01ed-4d7c-9fb8-576dd2866821" path="/var/lib/kubelet/pods/4075bf4d-01ed-4d7c-9fb8-576dd2866821/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.243938 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" path="/var/lib/kubelet/pods/4aeeba9b-9894-4293-a673-7e88c6bc80f5/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.245078 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" path="/var/lib/kubelet/pods/4b7f2c05-93d2-4139-95ea-c36bffcecdb8/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.245721 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c25a6d1-ee44-438c-a16d-e6952cfda420" path="/var/lib/kubelet/pods/4c25a6d1-ee44-438c-a16d-e6952cfda420/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.246327 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fad8d4a-3435-4647-b10d-51951c64e49a" path="/var/lib/kubelet/pods/4fad8d4a-3435-4647-b10d-51951c64e49a/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.247156 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bfee301-605c-4fb8-992e-8911bfc7206f" path="/var/lib/kubelet/pods/5bfee301-605c-4fb8-992e-8911bfc7206f/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.247612 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" path="/var/lib/kubelet/pods/67000f4b-b955-44d7-ad37-3d552d8efd88/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.248161 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" path="/var/lib/kubelet/pods/7b604164-1660-4f3a-929b-cf6f59b66823/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.249049 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="970e7515-bd36-45d9-befb-a0597cc72cbb" path="/var/lib/kubelet/pods/970e7515-bd36-45d9-befb-a0597cc72cbb/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.249672 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" path="/var/lib/kubelet/pods/a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.250281 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" path="/var/lib/kubelet/pods/a9adb465-b660-4d2c-bae8-f0b8024761b8/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.251247 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" path="/var/lib/kubelet/pods/ab6c469a-3a29-4fdb-b0ce-5671db6645d2/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.251935 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf293818-85fb-46dc-8cee-ba8eca827bb7" path="/var/lib/kubelet/pods/cf293818-85fb-46dc-8cee-ba8eca827bb7/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.252943 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" path="/var/lib/kubelet/pods/dcf4b178-75ee-47fb-81e4-e6b277eb1790/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.253537 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" path="/var/lib/kubelet/pods/df85b9e1-2830-40c3-bbea-555bcddf5545/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.254116 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" path="/var/lib/kubelet/pods/f2d4fc78-0ea7-49fe-9da3-caed0438e9f8/volumes" Oct 10 15:09:14 crc kubenswrapper[4788]: I1010 15:09:14.874540 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.179:8776/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 15:09:15 crc kubenswrapper[4788]: I1010 15:09:15.790109 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 15:09:15 crc kubenswrapper[4788]: I1010 15:09:15.790239 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 10 15:09:18 crc kubenswrapper[4788]: E1010 15:09:18.851597 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:18 crc kubenswrapper[4788]: E1010 15:09:18.852518 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:18 crc kubenswrapper[4788]: E1010 15:09:18.852919 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:18 crc kubenswrapper[4788]: E1010 15:09:18.853013 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:18 crc kubenswrapper[4788]: E1010 15:09:18.853835 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:18 crc kubenswrapper[4788]: E1010 15:09:18.855349 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:18 crc kubenswrapper[4788]: E1010 15:09:18.857180 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:18 crc kubenswrapper[4788]: E1010 15:09:18.857236 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.120854 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vz494"] Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121699 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="ceilometer-notification-agent" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121717 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="ceilometer-notification-agent" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121735 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121743 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121755 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121763 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-api" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121778 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerName="galera" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121786 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerName="galera" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121798 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121807 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121820 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerName="mysql-bootstrap" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121829 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerName="mysql-bootstrap" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121840 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121847 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121862 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4075bf4d-01ed-4d7c-9fb8-576dd2866821" containerName="memcached" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121869 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4075bf4d-01ed-4d7c-9fb8-576dd2866821" containerName="memcached" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121878 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerName="rabbitmq" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121885 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerName="rabbitmq" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121896 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerName="glance-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121905 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerName="glance-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121924 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="ceilometer-central-agent" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121931 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="ceilometer-central-agent" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121941 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerName="setup-container" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121948 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerName="setup-container" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121958 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-metadata" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121966 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-metadata" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121982 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.121989 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-log" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.121997 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerName="glance-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122004 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerName="glance-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122017 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerName="rabbitmq" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122026 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerName="rabbitmq" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122039 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerName="setup-container" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122047 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerName="setup-container" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122061 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fad8d4a-3435-4647-b10d-51951c64e49a" containerName="nova-scheduler-scheduler" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122068 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fad8d4a-3435-4647-b10d-51951c64e49a" containerName="nova-scheduler-scheduler" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122080 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerName="glance-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122088 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerName="glance-log" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122098 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="proxy-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122105 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="proxy-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122119 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122128 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122163 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="970e7515-bd36-45d9-befb-a0597cc72cbb" containerName="keystone-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122173 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="970e7515-bd36-45d9-befb-a0597cc72cbb" containerName="keystone-api" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122180 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122187 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-api" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122199 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerName="glance-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122207 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerName="glance-log" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122218 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf293818-85fb-46dc-8cee-ba8eca827bb7" containerName="nova-cell0-conductor-conductor" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122227 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf293818-85fb-46dc-8cee-ba8eca827bb7" containerName="nova-cell0-conductor-conductor" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122238 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="ovn-northd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122245 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="ovn-northd" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122260 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bfee301-605c-4fb8-992e-8911bfc7206f" containerName="mariadb-account-delete" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122267 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bfee301-605c-4fb8-992e-8911bfc7206f" containerName="mariadb-account-delete" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122274 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122281 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122294 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02360792-0295-42be-a94c-de0e7d2aa376" containerName="nova-cell1-conductor-conductor" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122301 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="02360792-0295-42be-a94c-de0e7d2aa376" containerName="nova-cell1-conductor-conductor" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122314 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122321 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-log" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122332 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="sg-core" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122341 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="sg-core" Oct 10 15:09:21 crc kubenswrapper[4788]: E1010 15:09:21.122354 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="openstack-network-exporter" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122360 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="openstack-network-exporter" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122527 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c25a6d1-ee44-438c-a16d-e6952cfda420" containerName="galera" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122538 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5b7cf2d-429d-42b1-8e42-a7fbdf012c2c" containerName="rabbitmq" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122552 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-metadata" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122563 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122576 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fad8d4a-3435-4647-b10d-51951c64e49a" containerName="nova-scheduler-scheduler" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122590 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcf4b178-75ee-47fb-81e4-e6b277eb1790" containerName="rabbitmq" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122599 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4075bf4d-01ed-4d7c-9fb8-576dd2866821" containerName="memcached" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122610 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="67000f4b-b955-44d7-ad37-3d552d8efd88" containerName="placement-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122624 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122636 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="sg-core" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122649 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerName="glance-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122660 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="ceilometer-notification-agent" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122670 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="ovn-northd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122684 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122696 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="ceilometer-central-agent" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122707 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerName="glance-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122719 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab6c469a-3a29-4fdb-b0ce-5671db6645d2" containerName="cinder-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122731 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bfee301-605c-4fb8-992e-8911bfc7206f" containerName="mariadb-account-delete" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122748 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9adb465-b660-4d2c-bae8-f0b8024761b8" containerName="openstack-network-exporter" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122760 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf293818-85fb-46dc-8cee-ba8eca827bb7" containerName="nova-cell0-conductor-conductor" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122772 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="02360792-0295-42be-a94c-de0e7d2aa376" containerName="nova-cell1-conductor-conductor" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122785 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="970e7515-bd36-45d9-befb-a0597cc72cbb" containerName="keystone-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122795 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122807 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2166478-a6f0-421b-9ea7-17f60ee9605b" containerName="glance-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122819 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b604164-1660-4f3a-929b-cf6f59b66823" containerName="barbican-api" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122831 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b7f2c05-93d2-4139-95ea-c36bffcecdb8" containerName="nova-api-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122840 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aeeba9b-9894-4293-a673-7e88c6bc80f5" containerName="proxy-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122847 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2d4fc78-0ea7-49fe-9da3-caed0438e9f8" containerName="nova-metadata-log" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.122857 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="df85b9e1-2830-40c3-bbea-555bcddf5545" containerName="glance-httpd" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.124315 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.159592 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vz494"] Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.222193 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xrfr\" (UniqueName: \"kubernetes.io/projected/d2559563-7735-4b3d-bc56-7cdc0cc11be0-kube-api-access-7xrfr\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.222265 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-utilities\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.222592 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-catalog-content\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.324502 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xrfr\" (UniqueName: \"kubernetes.io/projected/d2559563-7735-4b3d-bc56-7cdc0cc11be0-kube-api-access-7xrfr\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.324567 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-utilities\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.324662 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-catalog-content\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.325114 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-utilities\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.325423 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-catalog-content\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.374620 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xrfr\" (UniqueName: \"kubernetes.io/projected/d2559563-7735-4b3d-bc56-7cdc0cc11be0-kube-api-access-7xrfr\") pod \"redhat-operators-vz494\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.445626 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:21 crc kubenswrapper[4788]: I1010 15:09:21.949593 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vz494"] Oct 10 15:09:22 crc kubenswrapper[4788]: I1010 15:09:22.318550 4788 generic.go:334] "Generic (PLEG): container finished" podID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerID="29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d" exitCode=0 Oct 10 15:09:22 crc kubenswrapper[4788]: I1010 15:09:22.318593 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz494" event={"ID":"d2559563-7735-4b3d-bc56-7cdc0cc11be0","Type":"ContainerDied","Data":"29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d"} Oct 10 15:09:22 crc kubenswrapper[4788]: I1010 15:09:22.318620 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz494" event={"ID":"d2559563-7735-4b3d-bc56-7cdc0cc11be0","Type":"ContainerStarted","Data":"dffa4f756faf132c578abb3cb7152216023e40d910481515ae7b365e77ac2481"} Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.328967 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz494" event={"ID":"d2559563-7735-4b3d-bc56-7cdc0cc11be0","Type":"ContainerStarted","Data":"0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956"} Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.332013 4788 generic.go:334] "Generic (PLEG): container finished" podID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerID="2861b006539119d480308e351b33bc584593a678f96dce7d7c0ab944bb039ad4" exitCode=0 Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.332067 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-995b87f75-ghpz6" event={"ID":"e96b96e3-982f-4e27-97a6-4e076fc8ec40","Type":"ContainerDied","Data":"2861b006539119d480308e351b33bc584593a678f96dce7d7c0ab944bb039ad4"} Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.732708 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.760134 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-httpd-config\") pod \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.760227 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-ovndb-tls-certs\") pod \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.760293 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-public-tls-certs\") pod \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.760340 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-combined-ca-bundle\") pod \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.760368 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxfb6\" (UniqueName: \"kubernetes.io/projected/e96b96e3-982f-4e27-97a6-4e076fc8ec40-kube-api-access-lxfb6\") pod \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.760409 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-internal-tls-certs\") pod \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.760433 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-config\") pod \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\" (UID: \"e96b96e3-982f-4e27-97a6-4e076fc8ec40\") " Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.765281 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e96b96e3-982f-4e27-97a6-4e076fc8ec40" (UID: "e96b96e3-982f-4e27-97a6-4e076fc8ec40"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.765881 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e96b96e3-982f-4e27-97a6-4e076fc8ec40-kube-api-access-lxfb6" (OuterVolumeSpecName: "kube-api-access-lxfb6") pod "e96b96e3-982f-4e27-97a6-4e076fc8ec40" (UID: "e96b96e3-982f-4e27-97a6-4e076fc8ec40"). InnerVolumeSpecName "kube-api-access-lxfb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.809461 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e96b96e3-982f-4e27-97a6-4e076fc8ec40" (UID: "e96b96e3-982f-4e27-97a6-4e076fc8ec40"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.815517 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e96b96e3-982f-4e27-97a6-4e076fc8ec40" (UID: "e96b96e3-982f-4e27-97a6-4e076fc8ec40"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.822687 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e96b96e3-982f-4e27-97a6-4e076fc8ec40" (UID: "e96b96e3-982f-4e27-97a6-4e076fc8ec40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.823454 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e96b96e3-982f-4e27-97a6-4e076fc8ec40" (UID: "e96b96e3-982f-4e27-97a6-4e076fc8ec40"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.823867 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-config" (OuterVolumeSpecName: "config") pod "e96b96e3-982f-4e27-97a6-4e076fc8ec40" (UID: "e96b96e3-982f-4e27-97a6-4e076fc8ec40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:09:23 crc kubenswrapper[4788]: E1010 15:09:23.851808 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:23 crc kubenswrapper[4788]: E1010 15:09:23.852367 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:23 crc kubenswrapper[4788]: E1010 15:09:23.852619 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:23 crc kubenswrapper[4788]: E1010 15:09:23.852661 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:23 crc kubenswrapper[4788]: E1010 15:09:23.852870 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:23 crc kubenswrapper[4788]: E1010 15:09:23.855963 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:23 crc kubenswrapper[4788]: E1010 15:09:23.857953 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:23 crc kubenswrapper[4788]: E1010 15:09:23.857991 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.861776 4788 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.861807 4788 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.861819 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.861834 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxfb6\" (UniqueName: \"kubernetes.io/projected/e96b96e3-982f-4e27-97a6-4e076fc8ec40-kube-api-access-lxfb6\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.861847 4788 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.861858 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:23 crc kubenswrapper[4788]: I1010 15:09:23.861870 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e96b96e3-982f-4e27-97a6-4e076fc8ec40-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:24 crc kubenswrapper[4788]: I1010 15:09:24.344502 4788 generic.go:334] "Generic (PLEG): container finished" podID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerID="0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956" exitCode=0 Oct 10 15:09:24 crc kubenswrapper[4788]: I1010 15:09:24.344549 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz494" event={"ID":"d2559563-7735-4b3d-bc56-7cdc0cc11be0","Type":"ContainerDied","Data":"0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956"} Oct 10 15:09:24 crc kubenswrapper[4788]: I1010 15:09:24.347011 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-995b87f75-ghpz6" event={"ID":"e96b96e3-982f-4e27-97a6-4e076fc8ec40","Type":"ContainerDied","Data":"45f023f4d80ddb824b612b101bda21e1b042bacbc9a3fc698146cc9b7f763e15"} Oct 10 15:09:24 crc kubenswrapper[4788]: I1010 15:09:24.347062 4788 scope.go:117] "RemoveContainer" containerID="5699636a0ee4d24f0612d524e20711c956b0d7f42b60edadc77e819e598ab333" Oct 10 15:09:24 crc kubenswrapper[4788]: I1010 15:09:24.347212 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-995b87f75-ghpz6" Oct 10 15:09:24 crc kubenswrapper[4788]: I1010 15:09:24.383368 4788 scope.go:117] "RemoveContainer" containerID="2861b006539119d480308e351b33bc584593a678f96dce7d7c0ab944bb039ad4" Oct 10 15:09:24 crc kubenswrapper[4788]: I1010 15:09:24.386061 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-995b87f75-ghpz6"] Oct 10 15:09:24 crc kubenswrapper[4788]: I1010 15:09:24.391517 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-995b87f75-ghpz6"] Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.359566 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz494" event={"ID":"d2559563-7735-4b3d-bc56-7cdc0cc11be0","Type":"ContainerStarted","Data":"b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35"} Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.383790 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vz494" podStartSLOduration=1.802890599 podStartE2EDuration="4.383770085s" podCreationTimestamp="2025-10-10 15:09:21 +0000 UTC" firstStartedPulling="2025-10-10 15:09:22.321850936 +0000 UTC m=+1464.771566484" lastFinishedPulling="2025-10-10 15:09:24.902730412 +0000 UTC m=+1467.352445970" observedRunningTime="2025-10-10 15:09:25.38097975 +0000 UTC m=+1467.830695318" watchObservedRunningTime="2025-10-10 15:09:25.383770085 +0000 UTC m=+1467.833485633" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.870915 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mjcpf"] Oct 10 15:09:25 crc kubenswrapper[4788]: E1010 15:09:25.871515 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerName="neutron-httpd" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.871599 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerName="neutron-httpd" Oct 10 15:09:25 crc kubenswrapper[4788]: E1010 15:09:25.871697 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerName="neutron-api" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.871761 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerName="neutron-api" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.871993 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerName="neutron-api" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.872064 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" containerName="neutron-httpd" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.877608 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.897357 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mjcpf"] Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.996337 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-catalog-content\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.996415 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4d2z\" (UniqueName: \"kubernetes.io/projected/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-kube-api-access-g4d2z\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:25 crc kubenswrapper[4788]: I1010 15:09:25.996444 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-utilities\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.097867 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-catalog-content\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.097952 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4d2z\" (UniqueName: \"kubernetes.io/projected/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-kube-api-access-g4d2z\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.097981 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-utilities\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.098510 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-catalog-content\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.098556 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-utilities\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.121216 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4d2z\" (UniqueName: \"kubernetes.io/projected/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-kube-api-access-g4d2z\") pod \"certified-operators-mjcpf\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.212729 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.246095 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e96b96e3-982f-4e27-97a6-4e076fc8ec40" path="/var/lib/kubelet/pods/e96b96e3-982f-4e27-97a6-4e076fc8ec40/volumes" Oct 10 15:09:26 crc kubenswrapper[4788]: I1010 15:09:26.514914 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mjcpf"] Oct 10 15:09:27 crc kubenswrapper[4788]: I1010 15:09:27.393352 4788 generic.go:334] "Generic (PLEG): container finished" podID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerID="3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9" exitCode=0 Oct 10 15:09:27 crc kubenswrapper[4788]: I1010 15:09:27.393421 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mjcpf" event={"ID":"5bc35bf1-3af2-47a8-b21b-47f232ce9c38","Type":"ContainerDied","Data":"3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9"} Oct 10 15:09:27 crc kubenswrapper[4788]: I1010 15:09:27.394224 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mjcpf" event={"ID":"5bc35bf1-3af2-47a8-b21b-47f232ce9c38","Type":"ContainerStarted","Data":"b74d71e2d47bece0247db7a15c4785afd6c9d9c6906ea5e2f3636d10894a62a6"} Oct 10 15:09:28 crc kubenswrapper[4788]: E1010 15:09:28.850947 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:28 crc kubenswrapper[4788]: E1010 15:09:28.851702 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:28 crc kubenswrapper[4788]: E1010 15:09:28.852417 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:28 crc kubenswrapper[4788]: E1010 15:09:28.852471 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:28 crc kubenswrapper[4788]: E1010 15:09:28.853682 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:28 crc kubenswrapper[4788]: E1010 15:09:28.856136 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:28 crc kubenswrapper[4788]: E1010 15:09:28.858914 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:28 crc kubenswrapper[4788]: E1010 15:09:28.859007 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:29 crc kubenswrapper[4788]: I1010 15:09:29.406242 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:09:29 crc kubenswrapper[4788]: I1010 15:09:29.406323 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:09:29 crc kubenswrapper[4788]: I1010 15:09:29.406391 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:09:29 crc kubenswrapper[4788]: I1010 15:09:29.407395 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:09:29 crc kubenswrapper[4788]: I1010 15:09:29.407494 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" gracePeriod=600 Oct 10 15:09:29 crc kubenswrapper[4788]: I1010 15:09:29.417161 4788 generic.go:334] "Generic (PLEG): container finished" podID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerID="59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0" exitCode=0 Oct 10 15:09:29 crc kubenswrapper[4788]: I1010 15:09:29.417223 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mjcpf" event={"ID":"5bc35bf1-3af2-47a8-b21b-47f232ce9c38","Type":"ContainerDied","Data":"59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0"} Oct 10 15:09:29 crc kubenswrapper[4788]: E1010 15:09:29.561516 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:09:30 crc kubenswrapper[4788]: I1010 15:09:30.436446 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" exitCode=0 Oct 10 15:09:30 crc kubenswrapper[4788]: I1010 15:09:30.436505 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782"} Oct 10 15:09:30 crc kubenswrapper[4788]: I1010 15:09:30.436558 4788 scope.go:117] "RemoveContainer" containerID="8e260922adcdc34acddb0d03ca667c174167bee6681d51128f1688ce0f11d52b" Oct 10 15:09:30 crc kubenswrapper[4788]: I1010 15:09:30.437452 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:09:30 crc kubenswrapper[4788]: E1010 15:09:30.437978 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:09:31 crc kubenswrapper[4788]: I1010 15:09:31.446338 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:31 crc kubenswrapper[4788]: I1010 15:09:31.446485 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:31 crc kubenswrapper[4788]: I1010 15:09:31.525754 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:32 crc kubenswrapper[4788]: I1010 15:09:32.549363 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:32 crc kubenswrapper[4788]: I1010 15:09:32.662200 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vz494"] Oct 10 15:09:33 crc kubenswrapper[4788]: E1010 15:09:33.853953 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:33 crc kubenswrapper[4788]: E1010 15:09:33.856067 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:33 crc kubenswrapper[4788]: E1010 15:09:33.857242 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:33 crc kubenswrapper[4788]: E1010 15:09:33.857746 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:33 crc kubenswrapper[4788]: E1010 15:09:33.858593 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 10 15:09:33 crc kubenswrapper[4788]: E1010 15:09:33.858737 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:33 crc kubenswrapper[4788]: E1010 15:09:33.860401 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 10 15:09:33 crc kubenswrapper[4788]: E1010 15:09:33.860487 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-wpgvn" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.507416 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mjcpf" event={"ID":"5bc35bf1-3af2-47a8-b21b-47f232ce9c38","Type":"ContainerStarted","Data":"a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270"} Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.509987 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wpgvn_6627a65b-765d-430c-99f5-eddd285d4289/ovs-vswitchd/0.log" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.511738 4788 generic.go:334] "Generic (PLEG): container finished" podID="6627a65b-765d-430c-99f5-eddd285d4289" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" exitCode=137 Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.511899 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wpgvn" event={"ID":"6627a65b-765d-430c-99f5-eddd285d4289","Type":"ContainerDied","Data":"559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9"} Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.512176 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vz494" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerName="registry-server" containerID="cri-o://b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35" gracePeriod=2 Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.543972 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mjcpf" podStartSLOduration=2.7742645809999997 podStartE2EDuration="9.543947903s" podCreationTimestamp="2025-10-10 15:09:25 +0000 UTC" firstStartedPulling="2025-10-10 15:09:27.395178696 +0000 UTC m=+1469.844894244" lastFinishedPulling="2025-10-10 15:09:34.164862008 +0000 UTC m=+1476.614577566" observedRunningTime="2025-10-10 15:09:34.539694959 +0000 UTC m=+1476.989410537" watchObservedRunningTime="2025-10-10 15:09:34.543947903 +0000 UTC m=+1476.993663461" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.747576 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wpgvn_6627a65b-765d-430c-99f5-eddd285d4289/ovs-vswitchd/0.log" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.748713 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.866185 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-etc-ovs\") pod \"6627a65b-765d-430c-99f5-eddd285d4289\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.866276 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-run\") pod \"6627a65b-765d-430c-99f5-eddd285d4289\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.866353 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-log\") pod \"6627a65b-765d-430c-99f5-eddd285d4289\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.866458 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6627a65b-765d-430c-99f5-eddd285d4289-scripts\") pod \"6627a65b-765d-430c-99f5-eddd285d4289\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.866526 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-lib\") pod \"6627a65b-765d-430c-99f5-eddd285d4289\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.866599 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7jns\" (UniqueName: \"kubernetes.io/projected/6627a65b-765d-430c-99f5-eddd285d4289-kube-api-access-n7jns\") pod \"6627a65b-765d-430c-99f5-eddd285d4289\" (UID: \"6627a65b-765d-430c-99f5-eddd285d4289\") " Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.874847 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6627a65b-765d-430c-99f5-eddd285d4289-kube-api-access-n7jns" (OuterVolumeSpecName: "kube-api-access-n7jns") pod "6627a65b-765d-430c-99f5-eddd285d4289" (UID: "6627a65b-765d-430c-99f5-eddd285d4289"). InnerVolumeSpecName "kube-api-access-n7jns". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.874895 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-run" (OuterVolumeSpecName: "var-run") pod "6627a65b-765d-430c-99f5-eddd285d4289" (UID: "6627a65b-765d-430c-99f5-eddd285d4289"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.874936 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "6627a65b-765d-430c-99f5-eddd285d4289" (UID: "6627a65b-765d-430c-99f5-eddd285d4289"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.875064 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-lib" (OuterVolumeSpecName: "var-lib") pod "6627a65b-765d-430c-99f5-eddd285d4289" (UID: "6627a65b-765d-430c-99f5-eddd285d4289"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.875893 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6627a65b-765d-430c-99f5-eddd285d4289-scripts" (OuterVolumeSpecName: "scripts") pod "6627a65b-765d-430c-99f5-eddd285d4289" (UID: "6627a65b-765d-430c-99f5-eddd285d4289"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.874845 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-log" (OuterVolumeSpecName: "var-log") pod "6627a65b-765d-430c-99f5-eddd285d4289" (UID: "6627a65b-765d-430c-99f5-eddd285d4289"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.954693 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.968967 4788 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-lib\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.969007 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7jns\" (UniqueName: \"kubernetes.io/projected/6627a65b-765d-430c-99f5-eddd285d4289-kube-api-access-n7jns\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.969020 4788 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.969032 4788 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-run\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.969043 4788 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6627a65b-765d-430c-99f5-eddd285d4289-var-log\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:34 crc kubenswrapper[4788]: I1010 15:09:34.969057 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6627a65b-765d-430c-99f5-eddd285d4289-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.070944 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-catalog-content\") pod \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.071064 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-utilities\") pod \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.071196 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xrfr\" (UniqueName: \"kubernetes.io/projected/d2559563-7735-4b3d-bc56-7cdc0cc11be0-kube-api-access-7xrfr\") pod \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\" (UID: \"d2559563-7735-4b3d-bc56-7cdc0cc11be0\") " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.072015 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-utilities" (OuterVolumeSpecName: "utilities") pod "d2559563-7735-4b3d-bc56-7cdc0cc11be0" (UID: "d2559563-7735-4b3d-bc56-7cdc0cc11be0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.075068 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2559563-7735-4b3d-bc56-7cdc0cc11be0-kube-api-access-7xrfr" (OuterVolumeSpecName: "kube-api-access-7xrfr") pod "d2559563-7735-4b3d-bc56-7cdc0cc11be0" (UID: "d2559563-7735-4b3d-bc56-7cdc0cc11be0"). InnerVolumeSpecName "kube-api-access-7xrfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.156992 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2559563-7735-4b3d-bc56-7cdc0cc11be0" (UID: "d2559563-7735-4b3d-bc56-7cdc0cc11be0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.173569 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xrfr\" (UniqueName: \"kubernetes.io/projected/d2559563-7735-4b3d-bc56-7cdc0cc11be0-kube-api-access-7xrfr\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.173622 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.173670 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2559563-7735-4b3d-bc56-7cdc0cc11be0-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.527736 4788 generic.go:334] "Generic (PLEG): container finished" podID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerID="b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35" exitCode=0 Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.527836 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz494" event={"ID":"d2559563-7735-4b3d-bc56-7cdc0cc11be0","Type":"ContainerDied","Data":"b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35"} Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.527877 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz494" event={"ID":"d2559563-7735-4b3d-bc56-7cdc0cc11be0","Type":"ContainerDied","Data":"dffa4f756faf132c578abb3cb7152216023e40d910481515ae7b365e77ac2481"} Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.527918 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vz494" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.527922 4788 scope.go:117] "RemoveContainer" containerID="b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.533896 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wpgvn_6627a65b-765d-430c-99f5-eddd285d4289/ovs-vswitchd/0.log" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.534720 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wpgvn" event={"ID":"6627a65b-765d-430c-99f5-eddd285d4289","Type":"ContainerDied","Data":"e1ee3a66537f2628e8091a67b796dae5615ec3641c021bf5c3914a84f173fe37"} Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.534794 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-wpgvn" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.557528 4788 generic.go:334] "Generic (PLEG): container finished" podID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerID="ea921c62533da26ac37b94fc9c80bfc74b41e75cddc9adab1f4225855ea56a7d" exitCode=137 Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.558677 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"ea921c62533da26ac37b94fc9c80bfc74b41e75cddc9adab1f4225855ea56a7d"} Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.578262 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.580553 4788 scope.go:117] "RemoveContainer" containerID="0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.595931 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-wpgvn"] Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.609794 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-wpgvn"] Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.618380 4788 scope.go:117] "RemoveContainer" containerID="29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.620340 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vz494"] Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.627202 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vz494"] Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.644697 4788 scope.go:117] "RemoveContainer" containerID="b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35" Oct 10 15:09:35 crc kubenswrapper[4788]: E1010 15:09:35.645107 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35\": container with ID starting with b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35 not found: ID does not exist" containerID="b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.645156 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35"} err="failed to get container status \"b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35\": rpc error: code = NotFound desc = could not find container \"b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35\": container with ID starting with b17221790493e141141d73502202a669327625836382981b3e305c262bf48a35 not found: ID does not exist" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.645190 4788 scope.go:117] "RemoveContainer" containerID="0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956" Oct 10 15:09:35 crc kubenswrapper[4788]: E1010 15:09:35.645849 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956\": container with ID starting with 0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956 not found: ID does not exist" containerID="0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.645883 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956"} err="failed to get container status \"0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956\": rpc error: code = NotFound desc = could not find container \"0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956\": container with ID starting with 0f8817af2b0d91fa21e9b1a30d239aea2a835fbe6d0948c81ee67263379b4956 not found: ID does not exist" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.645899 4788 scope.go:117] "RemoveContainer" containerID="29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d" Oct 10 15:09:35 crc kubenswrapper[4788]: E1010 15:09:35.646284 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d\": container with ID starting with 29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d not found: ID does not exist" containerID="29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.646337 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d"} err="failed to get container status \"29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d\": rpc error: code = NotFound desc = could not find container \"29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d\": container with ID starting with 29fc14d46593be8fbf2457f60df7827a5392b807c5e0ded6d89913b0ed75188d not found: ID does not exist" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.646395 4788 scope.go:117] "RemoveContainer" containerID="559e0457c957099a79de13caa712a3ffa279be9a7bfda6f8f6fe1b6cfb8c8cd9" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.669829 4788 scope.go:117] "RemoveContainer" containerID="310cfd85a612b5e9482eb0926955357baa9456999571e1da15f4e0ef6f545b92" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.680898 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") pod \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.681347 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-lock\") pod \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.681502 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-cache\") pod \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.681756 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vl8d\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-kube-api-access-6vl8d\") pod \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.681871 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\" (UID: \"edada8e6-c183-4ef8-bcd5-f5c617e7615e\") " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.681887 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-lock" (OuterVolumeSpecName: "lock") pod "edada8e6-c183-4ef8-bcd5-f5c617e7615e" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.682155 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-cache" (OuterVolumeSpecName: "cache") pod "edada8e6-c183-4ef8-bcd5-f5c617e7615e" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.682496 4788 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-lock\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.682591 4788 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/edada8e6-c183-4ef8-bcd5-f5c617e7615e-cache\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.686216 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "swift") pod "edada8e6-c183-4ef8-bcd5-f5c617e7615e" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.687263 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-kube-api-access-6vl8d" (OuterVolumeSpecName: "kube-api-access-6vl8d") pod "edada8e6-c183-4ef8-bcd5-f5c617e7615e" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e"). InnerVolumeSpecName "kube-api-access-6vl8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.687545 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "edada8e6-c183-4ef8-bcd5-f5c617e7615e" (UID: "edada8e6-c183-4ef8-bcd5-f5c617e7615e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.690002 4788 scope.go:117] "RemoveContainer" containerID="1c8f0a077f21d24aafc30fb99c1d2ad328ddb6a3998fad99ac9bedd96c095da7" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.784393 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vl8d\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-kube-api-access-6vl8d\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.784667 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.784679 4788 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/edada8e6-c183-4ef8-bcd5-f5c617e7615e-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.799717 4788 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 10 15:09:35 crc kubenswrapper[4788]: I1010 15:09:35.888818 4788 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.213181 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.213247 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.244680 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6627a65b-765d-430c-99f5-eddd285d4289" path="/var/lib/kubelet/pods/6627a65b-765d-430c-99f5-eddd285d4289/volumes" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.246493 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" path="/var/lib/kubelet/pods/d2559563-7735-4b3d-bc56-7cdc0cc11be0/volumes" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.283043 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.588415 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"edada8e6-c183-4ef8-bcd5-f5c617e7615e","Type":"ContainerDied","Data":"a2b03249af073277654e5468956d1eebd61d5943a330291a73eb30edd0ffbd6e"} Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.588616 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.588659 4788 scope.go:117] "RemoveContainer" containerID="ea921c62533da26ac37b94fc9c80bfc74b41e75cddc9adab1f4225855ea56a7d" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.638833 4788 scope.go:117] "RemoveContainer" containerID="e1fa0e72188924318b11ba0840f43b627695d374bb14fe291577b87bd9109b71" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.642375 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.652307 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.673134 4788 scope.go:117] "RemoveContainer" containerID="21748426974e63b0540c88af385f3f29ae489df2ec5f695b422d31559e86f185" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.709587 4788 scope.go:117] "RemoveContainer" containerID="397c0d1e240027ba5e81338ba3e8dbb1354f4d23cae97d7d881c24cc8254d588" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.746762 4788 scope.go:117] "RemoveContainer" containerID="bb676839b319bffb1dc7ed545f7044cc6b58cf9440fc39529f2f530eaababd0f" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.783436 4788 scope.go:117] "RemoveContainer" containerID="292cc90a59ee5b84483d07d2af032969b7328c7cc913e39ba2daca604f9df39c" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.821307 4788 scope.go:117] "RemoveContainer" containerID="d37f41271c5034f6009e04dd6c3c22abb922d7b658aa8853581bf17fef626476" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.864587 4788 scope.go:117] "RemoveContainer" containerID="4e4b1fea33830faff4707c7c32b35feb21cbafff038fad525b1d34bd1ddc3463" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.900098 4788 scope.go:117] "RemoveContainer" containerID="da1d35d43d8e68922f14f6c617628576709461d2e0a8268d451f4cc8d694b90c" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.931820 4788 scope.go:117] "RemoveContainer" containerID="7e17d3abb98cf36468ce697828f20b0a33f4bdac07da3b2f7a9793d696601d10" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.962403 4788 scope.go:117] "RemoveContainer" containerID="642b90028a4eb24c71624969b7de73f0a0d672a8124912cb4fa0a195211e072e" Oct 10 15:09:36 crc kubenswrapper[4788]: I1010 15:09:36.989120 4788 scope.go:117] "RemoveContainer" containerID="3ffc6ad965d6e1a531e493d87c29df6c3edcf3b174067bab2eec1c073223fd6c" Oct 10 15:09:37 crc kubenswrapper[4788]: I1010 15:09:37.018554 4788 scope.go:117] "RemoveContainer" containerID="c68f1e8fc4064c0e04c7dd51691e00ff828f26e781caa02ae72f7ca08e509516" Oct 10 15:09:37 crc kubenswrapper[4788]: I1010 15:09:37.047066 4788 scope.go:117] "RemoveContainer" containerID="bf79745568579090730afa53a1af2d611537375a823410f2aefdb5bb4ae1de4c" Oct 10 15:09:37 crc kubenswrapper[4788]: I1010 15:09:37.071038 4788 scope.go:117] "RemoveContainer" containerID="3c7dd5fa50a9cff710f02fddbf4c3623aa2fc790f2bae6d263715c6a2a0be29a" Oct 10 15:09:38 crc kubenswrapper[4788]: I1010 15:09:38.248024 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" path="/var/lib/kubelet/pods/edada8e6-c183-4ef8-bcd5-f5c617e7615e/volumes" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.073651 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vqwms"] Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.073983 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server-init" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074000 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server-init" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074017 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-updater" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074023 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-updater" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074034 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="rsync" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074041 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="rsync" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074050 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-reaper" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074056 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-reaper" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074081 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074087 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074096 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerName="registry-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074102 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerName="registry-server" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074117 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074124 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-server" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074135 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074157 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-server" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074167 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="swift-recon-cron" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074174 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="swift-recon-cron" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074183 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074189 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074200 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-expirer" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074207 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-expirer" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074214 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerName="extract-utilities" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074221 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerName="extract-utilities" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074233 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-updater" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074239 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-updater" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074247 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074253 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074263 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074269 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074282 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074288 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074298 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074304 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-server" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074316 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerName="extract-content" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074321 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerName="extract-content" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074332 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074339 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074349 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074355 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:39 crc kubenswrapper[4788]: E1010 15:09:39.074364 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074371 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074518 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074529 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074540 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074551 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074562 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074569 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="swift-recon-cron" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074578 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovsdb-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074584 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="rsync" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074602 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6627a65b-765d-430c-99f5-eddd285d4289" containerName="ovs-vswitchd" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074609 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2559563-7735-4b3d-bc56-7cdc0cc11be0" containerName="registry-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074615 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-reaper" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074623 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-expirer" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074631 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074637 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-updater" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074645 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-updater" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074652 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="account-server" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074659 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="container-auditor" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.074667 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="edada8e6-c183-4ef8-bcd5-f5c617e7615e" containerName="object-replicator" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.075779 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.099685 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqwms"] Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.255367 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-utilities\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.255454 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fbw9\" (UniqueName: \"kubernetes.io/projected/d18bbfdd-7820-4782-9565-8609ea9c7292-kube-api-access-4fbw9\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.255553 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-catalog-content\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.356997 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-catalog-content\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.357110 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-utilities\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.357191 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fbw9\" (UniqueName: \"kubernetes.io/projected/d18bbfdd-7820-4782-9565-8609ea9c7292-kube-api-access-4fbw9\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.358093 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-catalog-content\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.358818 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-utilities\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.381613 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fbw9\" (UniqueName: \"kubernetes.io/projected/d18bbfdd-7820-4782-9565-8609ea9c7292-kube-api-access-4fbw9\") pod \"community-operators-vqwms\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.398096 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:39 crc kubenswrapper[4788]: I1010 15:09:39.932595 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqwms"] Oct 10 15:09:40 crc kubenswrapper[4788]: I1010 15:09:40.531847 4788 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podc35da072-521f-4266-9600-688bf5b49c1c"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podc35da072-521f-4266-9600-688bf5b49c1c] : Timed out while waiting for systemd to remove kubepods-besteffort-podc35da072_521f_4266_9600_688bf5b49c1c.slice" Oct 10 15:09:40 crc kubenswrapper[4788]: E1010 15:09:40.532316 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podc35da072-521f-4266-9600-688bf5b49c1c] : unable to destroy cgroup paths for cgroup [kubepods besteffort podc35da072-521f-4266-9600-688bf5b49c1c] : Timed out while waiting for systemd to remove kubepods-besteffort-podc35da072_521f_4266_9600_688bf5b49c1c.slice" pod="openstack/nova-cell1-novncproxy-0" podUID="c35da072-521f-4266-9600-688bf5b49c1c" Oct 10 15:09:40 crc kubenswrapper[4788]: I1010 15:09:40.674228 4788 generic.go:334] "Generic (PLEG): container finished" podID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerID="9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65" exitCode=0 Oct 10 15:09:40 crc kubenswrapper[4788]: I1010 15:09:40.674387 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqwms" event={"ID":"d18bbfdd-7820-4782-9565-8609ea9c7292","Type":"ContainerDied","Data":"9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65"} Oct 10 15:09:40 crc kubenswrapper[4788]: I1010 15:09:40.674429 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 15:09:40 crc kubenswrapper[4788]: I1010 15:09:40.674479 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqwms" event={"ID":"d18bbfdd-7820-4782-9565-8609ea9c7292","Type":"ContainerStarted","Data":"712a446fb8ed69f43ab9f8b0a9fd15b96a7d71205ffee5e7f64fb64da674b241"} Oct 10 15:09:40 crc kubenswrapper[4788]: I1010 15:09:40.742384 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:09:40 crc kubenswrapper[4788]: I1010 15:09:40.749590 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.243478 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c35da072-521f-4266-9600-688bf5b49c1c" path="/var/lib/kubelet/pods/c35da072-521f-4266-9600-688bf5b49c1c/volumes" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.321289 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0bb0b-account-delete-lcsdj" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.328064 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronab26-account-delete-g2dnr" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.337821 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi96b1-account-delete-pm4dv" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.393714 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance9145-account-delete-kblv8" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.418697 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vchst\" (UniqueName: \"kubernetes.io/projected/88f8c7ed-d478-4bae-ae23-4bffcb16abf6-kube-api-access-vchst\") pod \"88f8c7ed-d478-4bae-ae23-4bffcb16abf6\" (UID: \"88f8c7ed-d478-4bae-ae23-4bffcb16abf6\") " Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.419029 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnlg9\" (UniqueName: \"kubernetes.io/projected/92ccaefc-72b3-4e2f-90c0-394240ed2ab2-kube-api-access-jnlg9\") pod \"92ccaefc-72b3-4e2f-90c0-394240ed2ab2\" (UID: \"92ccaefc-72b3-4e2f-90c0-394240ed2ab2\") " Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.424275 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92ccaefc-72b3-4e2f-90c0-394240ed2ab2-kube-api-access-jnlg9" (OuterVolumeSpecName: "kube-api-access-jnlg9") pod "92ccaefc-72b3-4e2f-90c0-394240ed2ab2" (UID: "92ccaefc-72b3-4e2f-90c0-394240ed2ab2"). InnerVolumeSpecName "kube-api-access-jnlg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.424343 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88f8c7ed-d478-4bae-ae23-4bffcb16abf6-kube-api-access-vchst" (OuterVolumeSpecName: "kube-api-access-vchst") pod "88f8c7ed-d478-4bae-ae23-4bffcb16abf6" (UID: "88f8c7ed-d478-4bae-ae23-4bffcb16abf6"). InnerVolumeSpecName "kube-api-access-vchst". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.520885 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg9gc\" (UniqueName: \"kubernetes.io/projected/42d4825c-6e5c-4483-ae09-657903e6acb0-kube-api-access-kg9gc\") pod \"42d4825c-6e5c-4483-ae09-657903e6acb0\" (UID: \"42d4825c-6e5c-4483-ae09-657903e6acb0\") " Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.521027 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz8sx\" (UniqueName: \"kubernetes.io/projected/0d375458-0dc4-45a2-abf9-c7376510d2f6-kube-api-access-jz8sx\") pod \"0d375458-0dc4-45a2-abf9-c7376510d2f6\" (UID: \"0d375458-0dc4-45a2-abf9-c7376510d2f6\") " Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.521305 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vchst\" (UniqueName: \"kubernetes.io/projected/88f8c7ed-d478-4bae-ae23-4bffcb16abf6-kube-api-access-vchst\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.521323 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnlg9\" (UniqueName: \"kubernetes.io/projected/92ccaefc-72b3-4e2f-90c0-394240ed2ab2-kube-api-access-jnlg9\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.524884 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d4825c-6e5c-4483-ae09-657903e6acb0-kube-api-access-kg9gc" (OuterVolumeSpecName: "kube-api-access-kg9gc") pod "42d4825c-6e5c-4483-ae09-657903e6acb0" (UID: "42d4825c-6e5c-4483-ae09-657903e6acb0"). InnerVolumeSpecName "kube-api-access-kg9gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.526735 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d375458-0dc4-45a2-abf9-c7376510d2f6-kube-api-access-jz8sx" (OuterVolumeSpecName: "kube-api-access-jz8sx") pod "0d375458-0dc4-45a2-abf9-c7376510d2f6" (UID: "0d375458-0dc4-45a2-abf9-c7376510d2f6"). InnerVolumeSpecName "kube-api-access-jz8sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.623010 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz8sx\" (UniqueName: \"kubernetes.io/projected/0d375458-0dc4-45a2-abf9-c7376510d2f6-kube-api-access-jz8sx\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.623071 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg9gc\" (UniqueName: \"kubernetes.io/projected/42d4825c-6e5c-4483-ae09-657903e6acb0-kube-api-access-kg9gc\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.702374 4788 generic.go:334] "Generic (PLEG): container finished" podID="42d4825c-6e5c-4483-ae09-657903e6acb0" containerID="5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b" exitCode=137 Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.702412 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance9145-account-delete-kblv8" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.702446 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance9145-account-delete-kblv8" event={"ID":"42d4825c-6e5c-4483-ae09-657903e6acb0","Type":"ContainerDied","Data":"5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.703436 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance9145-account-delete-kblv8" event={"ID":"42d4825c-6e5c-4483-ae09-657903e6acb0","Type":"ContainerDied","Data":"dd80b547d77a987656b69fe0189950742a72e4c531b826b36706620344d598b4"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.703577 4788 scope.go:117] "RemoveContainer" containerID="5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.707055 4788 generic.go:334] "Generic (PLEG): container finished" podID="92ccaefc-72b3-4e2f-90c0-394240ed2ab2" containerID="39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb" exitCode=137 Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.707340 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0bb0b-account-delete-lcsdj" event={"ID":"92ccaefc-72b3-4e2f-90c0-394240ed2ab2","Type":"ContainerDied","Data":"39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.707370 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0bb0b-account-delete-lcsdj" event={"ID":"92ccaefc-72b3-4e2f-90c0-394240ed2ab2","Type":"ContainerDied","Data":"6da28bc2ea7d4101fea6c67bfcb3e698df4aa0c013b90d1e81ec7d3a016e1f77"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.707441 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0bb0b-account-delete-lcsdj" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.713659 4788 generic.go:334] "Generic (PLEG): container finished" podID="88f8c7ed-d478-4bae-ae23-4bffcb16abf6" containerID="d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259" exitCode=137 Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.713734 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronab26-account-delete-g2dnr" event={"ID":"88f8c7ed-d478-4bae-ae23-4bffcb16abf6","Type":"ContainerDied","Data":"d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.713768 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronab26-account-delete-g2dnr" event={"ID":"88f8c7ed-d478-4bae-ae23-4bffcb16abf6","Type":"ContainerDied","Data":"d8adf007843a2beec7317755d482e4c66746a4321f0e4281a315c76cdb5da46e"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.713826 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronab26-account-delete-g2dnr" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.718272 4788 generic.go:334] "Generic (PLEG): container finished" podID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerID="90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568" exitCode=0 Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.718395 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqwms" event={"ID":"d18bbfdd-7820-4782-9565-8609ea9c7292","Type":"ContainerDied","Data":"90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.720959 4788 generic.go:334] "Generic (PLEG): container finished" podID="0d375458-0dc4-45a2-abf9-c7376510d2f6" containerID="5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76" exitCode=137 Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.721006 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi96b1-account-delete-pm4dv" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.721008 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi96b1-account-delete-pm4dv" event={"ID":"0d375458-0dc4-45a2-abf9-c7376510d2f6","Type":"ContainerDied","Data":"5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.721045 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi96b1-account-delete-pm4dv" event={"ID":"0d375458-0dc4-45a2-abf9-c7376510d2f6","Type":"ContainerDied","Data":"f7417620e756c0bf0389f9a52443ea4e26c42c102ca08ae937e8e3232743307a"} Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.791056 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronab26-account-delete-g2dnr"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.791296 4788 scope.go:117] "RemoveContainer" containerID="5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b" Oct 10 15:09:42 crc kubenswrapper[4788]: E1010 15:09:42.792459 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b\": container with ID starting with 5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b not found: ID does not exist" containerID="5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.792552 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b"} err="failed to get container status \"5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b\": rpc error: code = NotFound desc = could not find container \"5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b\": container with ID starting with 5428a127d95640c174c4e49ae7ac255e8af08df497caaaaa0aece237bd0ea44b not found: ID does not exist" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.792613 4788 scope.go:117] "RemoveContainer" containerID="39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.804749 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronab26-account-delete-g2dnr"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.824135 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0bb0b-account-delete-lcsdj"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.831423 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0bb0b-account-delete-lcsdj"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.833309 4788 scope.go:117] "RemoveContainer" containerID="39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb" Oct 10 15:09:42 crc kubenswrapper[4788]: E1010 15:09:42.833735 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb\": container with ID starting with 39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb not found: ID does not exist" containerID="39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.833770 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb"} err="failed to get container status \"39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb\": rpc error: code = NotFound desc = could not find container \"39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb\": container with ID starting with 39b7c0ad50f39b03faed3e431dd720043f9e63e2513aa5479dd8492fea2322cb not found: ID does not exist" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.833796 4788 scope.go:117] "RemoveContainer" containerID="d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.839265 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi96b1-account-delete-pm4dv"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.846631 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi96b1-account-delete-pm4dv"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.872252 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance9145-account-delete-kblv8"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.872929 4788 scope.go:117] "RemoveContainer" containerID="d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259" Oct 10 15:09:42 crc kubenswrapper[4788]: E1010 15:09:42.873902 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259\": container with ID starting with d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259 not found: ID does not exist" containerID="d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.873967 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259"} err="failed to get container status \"d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259\": rpc error: code = NotFound desc = could not find container \"d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259\": container with ID starting with d6dd939fe2bcc04a1fcac119dd41b6728e774680b772daabaedb393c163db259 not found: ID does not exist" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.873999 4788 scope.go:117] "RemoveContainer" containerID="5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.882364 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance9145-account-delete-kblv8"] Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.897504 4788 scope.go:117] "RemoveContainer" containerID="5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76" Oct 10 15:09:42 crc kubenswrapper[4788]: E1010 15:09:42.897971 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76\": container with ID starting with 5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76 not found: ID does not exist" containerID="5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76" Oct 10 15:09:42 crc kubenswrapper[4788]: I1010 15:09:42.898027 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76"} err="failed to get container status \"5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76\": rpc error: code = NotFound desc = could not find container \"5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76\": container with ID starting with 5c417cb54c3269da468e7db1f2c026603985f49d996aa9decd963331d1025f76 not found: ID does not exist" Oct 10 15:09:43 crc kubenswrapper[4788]: I1010 15:09:43.233988 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:09:43 crc kubenswrapper[4788]: E1010 15:09:43.234671 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:09:43 crc kubenswrapper[4788]: I1010 15:09:43.739628 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqwms" event={"ID":"d18bbfdd-7820-4782-9565-8609ea9c7292","Type":"ContainerStarted","Data":"c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a"} Oct 10 15:09:43 crc kubenswrapper[4788]: I1010 15:09:43.764787 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vqwms" podStartSLOduration=2.317388986 podStartE2EDuration="4.764770546s" podCreationTimestamp="2025-10-10 15:09:39 +0000 UTC" firstStartedPulling="2025-10-10 15:09:40.678665565 +0000 UTC m=+1483.128381143" lastFinishedPulling="2025-10-10 15:09:43.126047125 +0000 UTC m=+1485.575762703" observedRunningTime="2025-10-10 15:09:43.762999739 +0000 UTC m=+1486.212715307" watchObservedRunningTime="2025-10-10 15:09:43.764770546 +0000 UTC m=+1486.214486094" Oct 10 15:09:44 crc kubenswrapper[4788]: I1010 15:09:44.253685 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d375458-0dc4-45a2-abf9-c7376510d2f6" path="/var/lib/kubelet/pods/0d375458-0dc4-45a2-abf9-c7376510d2f6/volumes" Oct 10 15:09:44 crc kubenswrapper[4788]: I1010 15:09:44.255502 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42d4825c-6e5c-4483-ae09-657903e6acb0" path="/var/lib/kubelet/pods/42d4825c-6e5c-4483-ae09-657903e6acb0/volumes" Oct 10 15:09:44 crc kubenswrapper[4788]: I1010 15:09:44.256471 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88f8c7ed-d478-4bae-ae23-4bffcb16abf6" path="/var/lib/kubelet/pods/88f8c7ed-d478-4bae-ae23-4bffcb16abf6/volumes" Oct 10 15:09:44 crc kubenswrapper[4788]: I1010 15:09:44.257574 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92ccaefc-72b3-4e2f-90c0-394240ed2ab2" path="/var/lib/kubelet/pods/92ccaefc-72b3-4e2f-90c0-394240ed2ab2/volumes" Oct 10 15:09:46 crc kubenswrapper[4788]: I1010 15:09:46.260611 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:46 crc kubenswrapper[4788]: I1010 15:09:46.312120 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mjcpf"] Oct 10 15:09:46 crc kubenswrapper[4788]: I1010 15:09:46.770574 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mjcpf" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerName="registry-server" containerID="cri-o://a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270" gracePeriod=2 Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.213866 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.297579 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-utilities\") pod \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.297683 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-catalog-content\") pod \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.297957 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4d2z\" (UniqueName: \"kubernetes.io/projected/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-kube-api-access-g4d2z\") pod \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\" (UID: \"5bc35bf1-3af2-47a8-b21b-47f232ce9c38\") " Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.299259 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-utilities" (OuterVolumeSpecName: "utilities") pod "5bc35bf1-3af2-47a8-b21b-47f232ce9c38" (UID: "5bc35bf1-3af2-47a8-b21b-47f232ce9c38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.303947 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.317201 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-kube-api-access-g4d2z" (OuterVolumeSpecName: "kube-api-access-g4d2z") pod "5bc35bf1-3af2-47a8-b21b-47f232ce9c38" (UID: "5bc35bf1-3af2-47a8-b21b-47f232ce9c38"). InnerVolumeSpecName "kube-api-access-g4d2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.365271 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bc35bf1-3af2-47a8-b21b-47f232ce9c38" (UID: "5bc35bf1-3af2-47a8-b21b-47f232ce9c38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.405613 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.405646 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4d2z\" (UniqueName: \"kubernetes.io/projected/5bc35bf1-3af2-47a8-b21b-47f232ce9c38-kube-api-access-g4d2z\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.782871 4788 generic.go:334] "Generic (PLEG): container finished" podID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerID="a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270" exitCode=0 Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.782929 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mjcpf" event={"ID":"5bc35bf1-3af2-47a8-b21b-47f232ce9c38","Type":"ContainerDied","Data":"a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270"} Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.782963 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mjcpf" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.782988 4788 scope.go:117] "RemoveContainer" containerID="a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.782969 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mjcpf" event={"ID":"5bc35bf1-3af2-47a8-b21b-47f232ce9c38","Type":"ContainerDied","Data":"b74d71e2d47bece0247db7a15c4785afd6c9d9c6906ea5e2f3636d10894a62a6"} Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.816057 4788 scope.go:117] "RemoveContainer" containerID="59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.826620 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mjcpf"] Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.834411 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mjcpf"] Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.855317 4788 scope.go:117] "RemoveContainer" containerID="3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.881169 4788 scope.go:117] "RemoveContainer" containerID="a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270" Oct 10 15:09:47 crc kubenswrapper[4788]: E1010 15:09:47.882033 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270\": container with ID starting with a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270 not found: ID does not exist" containerID="a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.882126 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270"} err="failed to get container status \"a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270\": rpc error: code = NotFound desc = could not find container \"a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270\": container with ID starting with a82ec18d347cffd4c3a4e2b5c98a3ef49957b1716ec75d33d627bc5d64142270 not found: ID does not exist" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.882273 4788 scope.go:117] "RemoveContainer" containerID="59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0" Oct 10 15:09:47 crc kubenswrapper[4788]: E1010 15:09:47.882746 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0\": container with ID starting with 59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0 not found: ID does not exist" containerID="59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.882793 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0"} err="failed to get container status \"59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0\": rpc error: code = NotFound desc = could not find container \"59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0\": container with ID starting with 59a90b9b82d3487e5b85835e9444f6d0db501e61f191f762dddbd001167cfae0 not found: ID does not exist" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.882825 4788 scope.go:117] "RemoveContainer" containerID="3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9" Oct 10 15:09:47 crc kubenswrapper[4788]: E1010 15:09:47.883287 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9\": container with ID starting with 3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9 not found: ID does not exist" containerID="3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9" Oct 10 15:09:47 crc kubenswrapper[4788]: I1010 15:09:47.883362 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9"} err="failed to get container status \"3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9\": rpc error: code = NotFound desc = could not find container \"3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9\": container with ID starting with 3a895c342bbb5268591703a66cca4fffdf7399783b899c8f350712706a7864c9 not found: ID does not exist" Oct 10 15:09:48 crc kubenswrapper[4788]: I1010 15:09:48.244714 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" path="/var/lib/kubelet/pods/5bc35bf1-3af2-47a8-b21b-47f232ce9c38/volumes" Oct 10 15:09:49 crc kubenswrapper[4788]: I1010 15:09:49.399046 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:49 crc kubenswrapper[4788]: I1010 15:09:49.399723 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:49 crc kubenswrapper[4788]: I1010 15:09:49.465518 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:49 crc kubenswrapper[4788]: I1010 15:09:49.884826 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:50 crc kubenswrapper[4788]: I1010 15:09:50.908226 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqwms"] Oct 10 15:09:51 crc kubenswrapper[4788]: I1010 15:09:51.843102 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vqwms" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerName="registry-server" containerID="cri-o://c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a" gracePeriod=2 Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.228849 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.301183 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-utilities\") pod \"d18bbfdd-7820-4782-9565-8609ea9c7292\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.301289 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fbw9\" (UniqueName: \"kubernetes.io/projected/d18bbfdd-7820-4782-9565-8609ea9c7292-kube-api-access-4fbw9\") pod \"d18bbfdd-7820-4782-9565-8609ea9c7292\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.301330 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-catalog-content\") pod \"d18bbfdd-7820-4782-9565-8609ea9c7292\" (UID: \"d18bbfdd-7820-4782-9565-8609ea9c7292\") " Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.302068 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-utilities" (OuterVolumeSpecName: "utilities") pod "d18bbfdd-7820-4782-9565-8609ea9c7292" (UID: "d18bbfdd-7820-4782-9565-8609ea9c7292"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.308340 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d18bbfdd-7820-4782-9565-8609ea9c7292-kube-api-access-4fbw9" (OuterVolumeSpecName: "kube-api-access-4fbw9") pod "d18bbfdd-7820-4782-9565-8609ea9c7292" (UID: "d18bbfdd-7820-4782-9565-8609ea9c7292"). InnerVolumeSpecName "kube-api-access-4fbw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.363218 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d18bbfdd-7820-4782-9565-8609ea9c7292" (UID: "d18bbfdd-7820-4782-9565-8609ea9c7292"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.402921 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.403125 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d18bbfdd-7820-4782-9565-8609ea9c7292-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.403229 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fbw9\" (UniqueName: \"kubernetes.io/projected/d18bbfdd-7820-4782-9565-8609ea9c7292-kube-api-access-4fbw9\") on node \"crc\" DevicePath \"\"" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.858060 4788 generic.go:334] "Generic (PLEG): container finished" podID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerID="c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a" exitCode=0 Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.858134 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqwms" event={"ID":"d18bbfdd-7820-4782-9565-8609ea9c7292","Type":"ContainerDied","Data":"c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a"} Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.858200 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqwms" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.858239 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqwms" event={"ID":"d18bbfdd-7820-4782-9565-8609ea9c7292","Type":"ContainerDied","Data":"712a446fb8ed69f43ab9f8b0a9fd15b96a7d71205ffee5e7f64fb64da674b241"} Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.858279 4788 scope.go:117] "RemoveContainer" containerID="c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.916534 4788 scope.go:117] "RemoveContainer" containerID="90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.920908 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqwms"] Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.929476 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vqwms"] Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.951852 4788 scope.go:117] "RemoveContainer" containerID="9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.994560 4788 scope.go:117] "RemoveContainer" containerID="c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a" Oct 10 15:09:52 crc kubenswrapper[4788]: E1010 15:09:52.995281 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a\": container with ID starting with c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a not found: ID does not exist" containerID="c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.995386 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a"} err="failed to get container status \"c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a\": rpc error: code = NotFound desc = could not find container \"c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a\": container with ID starting with c496ad537002cec79113448cae8b8a8e866718562d315e68c6fad0e4f1ddbc1a not found: ID does not exist" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.995454 4788 scope.go:117] "RemoveContainer" containerID="90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568" Oct 10 15:09:52 crc kubenswrapper[4788]: E1010 15:09:52.996006 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568\": container with ID starting with 90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568 not found: ID does not exist" containerID="90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.996073 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568"} err="failed to get container status \"90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568\": rpc error: code = NotFound desc = could not find container \"90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568\": container with ID starting with 90feb7e8dfdcf99fea4e62dbfb93d85eed100a6de090f7d650961561ea98e568 not found: ID does not exist" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.996109 4788 scope.go:117] "RemoveContainer" containerID="9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65" Oct 10 15:09:52 crc kubenswrapper[4788]: E1010 15:09:52.996515 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65\": container with ID starting with 9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65 not found: ID does not exist" containerID="9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65" Oct 10 15:09:52 crc kubenswrapper[4788]: I1010 15:09:52.996553 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65"} err="failed to get container status \"9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65\": rpc error: code = NotFound desc = could not find container \"9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65\": container with ID starting with 9e4582cc19995ff43bb3a68bdfea7a2da4354c223cec49b65170d87d90fe6c65 not found: ID does not exist" Oct 10 15:09:54 crc kubenswrapper[4788]: I1010 15:09:54.247247 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" path="/var/lib/kubelet/pods/d18bbfdd-7820-4782-9565-8609ea9c7292/volumes" Oct 10 15:09:57 crc kubenswrapper[4788]: I1010 15:09:57.233675 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:09:57 crc kubenswrapper[4788]: E1010 15:09:57.234096 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:10:06 crc kubenswrapper[4788]: I1010 15:10:06.824885 4788 scope.go:117] "RemoveContainer" containerID="6dfca80283e228ab25b825ce8120c260fb7c568dd80ba2c02f1a941cc12700a6" Oct 10 15:10:08 crc kubenswrapper[4788]: I1010 15:10:08.234383 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:10:08 crc kubenswrapper[4788]: E1010 15:10:08.234851 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:10:22 crc kubenswrapper[4788]: I1010 15:10:22.234317 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:10:22 crc kubenswrapper[4788]: E1010 15:10:22.234989 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:10:36 crc kubenswrapper[4788]: I1010 15:10:36.234709 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:10:36 crc kubenswrapper[4788]: E1010 15:10:36.235865 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:10:49 crc kubenswrapper[4788]: I1010 15:10:49.234604 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:10:49 crc kubenswrapper[4788]: E1010 15:10:49.235854 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:11:01 crc kubenswrapper[4788]: I1010 15:11:01.237587 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:11:01 crc kubenswrapper[4788]: E1010 15:11:01.238926 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.609726 4788 scope.go:117] "RemoveContainer" containerID="f4727ecaab5a16f71e927909b73eabb95375907d30bc07c7b99154a067738b3b" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.640745 4788 scope.go:117] "RemoveContainer" containerID="f2e78e25b19d6852b324ff847adff58351738286f1c416abcc7aab836feedf39" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.695710 4788 scope.go:117] "RemoveContainer" containerID="7e850bcecdc4369a4a000f3af74b9e10aa254c1fd8e0b4cd9c9bce7f62cee54e" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.737107 4788 scope.go:117] "RemoveContainer" containerID="4718475c27409de0e7a18cce2d0271c55efe267c0167c7e7a90691925b3c008d" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.764893 4788 scope.go:117] "RemoveContainer" containerID="93cdd8df755829312c77164115af990c8b6b186683f6a2fa4048af1e3e662e81" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.790523 4788 scope.go:117] "RemoveContainer" containerID="0e206d9712b2fcf1d859caed9a81a00ff94cd840d0e81d1c761455259fa5b1b2" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.814825 4788 scope.go:117] "RemoveContainer" containerID="5a5958beeb484fa10133ebb3bdf2fb0ffb0366c0052b30e8639799c67e07fcc9" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.836049 4788 scope.go:117] "RemoveContainer" containerID="87b5409f5a03eb92c52a01f84e674066fa474f0acba5914786ae0ae6c9f1504d" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.858657 4788 scope.go:117] "RemoveContainer" containerID="6d31303eab8e8df640fcd2caa8d708c4dba787f55933092d07939f7904769a25" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.880753 4788 scope.go:117] "RemoveContainer" containerID="c74f38ffd4f07c84655a1a8102dedf197116202fc557e371bc029b11a0688100" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.924542 4788 scope.go:117] "RemoveContainer" containerID="081f9a04373c71c05b380d5f33c6e373a025a23595bb2ef7f84dd4e4b4b298fe" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.966866 4788 scope.go:117] "RemoveContainer" containerID="3af2beb51561d3ef0612a1db993a775db975046929eb4d1f3edfa355eb9cecc3" Oct 10 15:11:07 crc kubenswrapper[4788]: I1010 15:11:07.992927 4788 scope.go:117] "RemoveContainer" containerID="f0a4c089fd7d0b89314e500a54e9e888fb4099f6bd2b5736fa18d7a297dc7323" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.613258 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7v7r7"] Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614307 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88f8c7ed-d478-4bae-ae23-4bffcb16abf6" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614324 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="88f8c7ed-d478-4bae-ae23-4bffcb16abf6" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614360 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerName="extract-content" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614369 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerName="extract-content" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614394 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerName="registry-server" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614402 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerName="registry-server" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614419 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerName="extract-content" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614427 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerName="extract-content" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614455 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d375458-0dc4-45a2-abf9-c7376510d2f6" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614464 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d375458-0dc4-45a2-abf9-c7376510d2f6" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614478 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d4825c-6e5c-4483-ae09-657903e6acb0" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614486 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d4825c-6e5c-4483-ae09-657903e6acb0" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614499 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerName="extract-utilities" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614507 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerName="extract-utilities" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614519 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerName="extract-utilities" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614527 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerName="extract-utilities" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614538 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerName="registry-server" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614546 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerName="registry-server" Oct 10 15:11:10 crc kubenswrapper[4788]: E1010 15:11:10.614560 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ccaefc-72b3-4e2f-90c0-394240ed2ab2" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614568 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ccaefc-72b3-4e2f-90c0-394240ed2ab2" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614748 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ccaefc-72b3-4e2f-90c0-394240ed2ab2" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614770 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d375458-0dc4-45a2-abf9-c7376510d2f6" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614794 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="88f8c7ed-d478-4bae-ae23-4bffcb16abf6" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614806 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d4825c-6e5c-4483-ae09-657903e6acb0" containerName="mariadb-account-delete" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614821 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc35bf1-3af2-47a8-b21b-47f232ce9c38" containerName="registry-server" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.614850 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d18bbfdd-7820-4782-9565-8609ea9c7292" containerName="registry-server" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.616201 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.674107 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7v7r7"] Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.714631 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6npf\" (UniqueName: \"kubernetes.io/projected/ff7d78a9-cb91-47e4-93bb-19d1b112a329-kube-api-access-b6npf\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.714678 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-catalog-content\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.714748 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-utilities\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.816545 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-utilities\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.816642 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6npf\" (UniqueName: \"kubernetes.io/projected/ff7d78a9-cb91-47e4-93bb-19d1b112a329-kube-api-access-b6npf\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.816673 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-catalog-content\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.817189 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-catalog-content\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.817227 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-utilities\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.842083 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6npf\" (UniqueName: \"kubernetes.io/projected/ff7d78a9-cb91-47e4-93bb-19d1b112a329-kube-api-access-b6npf\") pod \"redhat-marketplace-7v7r7\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:10 crc kubenswrapper[4788]: I1010 15:11:10.936353 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:11 crc kubenswrapper[4788]: I1010 15:11:11.411933 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7v7r7"] Oct 10 15:11:11 crc kubenswrapper[4788]: I1010 15:11:11.791853 4788 generic.go:334] "Generic (PLEG): container finished" podID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerID="42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0" exitCode=0 Oct 10 15:11:11 crc kubenswrapper[4788]: I1010 15:11:11.791928 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7v7r7" event={"ID":"ff7d78a9-cb91-47e4-93bb-19d1b112a329","Type":"ContainerDied","Data":"42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0"} Oct 10 15:11:11 crc kubenswrapper[4788]: I1010 15:11:11.792128 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7v7r7" event={"ID":"ff7d78a9-cb91-47e4-93bb-19d1b112a329","Type":"ContainerStarted","Data":"f57ed69b188915722fccb3cf89f03bd5814592fe908b9f55786dfad440985ae4"} Oct 10 15:11:12 crc kubenswrapper[4788]: I1010 15:11:12.801197 4788 generic.go:334] "Generic (PLEG): container finished" podID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerID="2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6" exitCode=0 Oct 10 15:11:12 crc kubenswrapper[4788]: I1010 15:11:12.801240 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7v7r7" event={"ID":"ff7d78a9-cb91-47e4-93bb-19d1b112a329","Type":"ContainerDied","Data":"2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6"} Oct 10 15:11:13 crc kubenswrapper[4788]: I1010 15:11:13.235172 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:11:13 crc kubenswrapper[4788]: E1010 15:11:13.235447 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:11:13 crc kubenswrapper[4788]: I1010 15:11:13.810429 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7v7r7" event={"ID":"ff7d78a9-cb91-47e4-93bb-19d1b112a329","Type":"ContainerStarted","Data":"42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882"} Oct 10 15:11:13 crc kubenswrapper[4788]: I1010 15:11:13.831642 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7v7r7" podStartSLOduration=2.425809761 podStartE2EDuration="3.831624153s" podCreationTimestamp="2025-10-10 15:11:10 +0000 UTC" firstStartedPulling="2025-10-10 15:11:11.793544993 +0000 UTC m=+1574.243260541" lastFinishedPulling="2025-10-10 15:11:13.199359365 +0000 UTC m=+1575.649074933" observedRunningTime="2025-10-10 15:11:13.829458945 +0000 UTC m=+1576.279174493" watchObservedRunningTime="2025-10-10 15:11:13.831624153 +0000 UTC m=+1576.281339711" Oct 10 15:11:20 crc kubenswrapper[4788]: I1010 15:11:20.937198 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:20 crc kubenswrapper[4788]: I1010 15:11:20.937867 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:20 crc kubenswrapper[4788]: I1010 15:11:20.991385 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:21 crc kubenswrapper[4788]: I1010 15:11:21.914832 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:21 crc kubenswrapper[4788]: I1010 15:11:21.966757 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7v7r7"] Oct 10 15:11:23 crc kubenswrapper[4788]: I1010 15:11:23.898661 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7v7r7" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerName="registry-server" containerID="cri-o://42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882" gracePeriod=2 Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.294627 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.434881 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6npf\" (UniqueName: \"kubernetes.io/projected/ff7d78a9-cb91-47e4-93bb-19d1b112a329-kube-api-access-b6npf\") pod \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.434994 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-catalog-content\") pod \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.435070 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-utilities\") pod \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\" (UID: \"ff7d78a9-cb91-47e4-93bb-19d1b112a329\") " Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.435917 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-utilities" (OuterVolumeSpecName: "utilities") pod "ff7d78a9-cb91-47e4-93bb-19d1b112a329" (UID: "ff7d78a9-cb91-47e4-93bb-19d1b112a329"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.441262 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7d78a9-cb91-47e4-93bb-19d1b112a329-kube-api-access-b6npf" (OuterVolumeSpecName: "kube-api-access-b6npf") pod "ff7d78a9-cb91-47e4-93bb-19d1b112a329" (UID: "ff7d78a9-cb91-47e4-93bb-19d1b112a329"). InnerVolumeSpecName "kube-api-access-b6npf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.449050 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff7d78a9-cb91-47e4-93bb-19d1b112a329" (UID: "ff7d78a9-cb91-47e4-93bb-19d1b112a329"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.536171 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.536225 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6npf\" (UniqueName: \"kubernetes.io/projected/ff7d78a9-cb91-47e4-93bb-19d1b112a329-kube-api-access-b6npf\") on node \"crc\" DevicePath \"\"" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.536238 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7d78a9-cb91-47e4-93bb-19d1b112a329-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.911661 4788 generic.go:334] "Generic (PLEG): container finished" podID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerID="42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882" exitCode=0 Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.911717 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7v7r7" event={"ID":"ff7d78a9-cb91-47e4-93bb-19d1b112a329","Type":"ContainerDied","Data":"42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882"} Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.911786 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7v7r7" event={"ID":"ff7d78a9-cb91-47e4-93bb-19d1b112a329","Type":"ContainerDied","Data":"f57ed69b188915722fccb3cf89f03bd5814592fe908b9f55786dfad440985ae4"} Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.911811 4788 scope.go:117] "RemoveContainer" containerID="42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.911808 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7v7r7" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.951446 4788 scope.go:117] "RemoveContainer" containerID="2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6" Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.955092 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7v7r7"] Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.962529 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7v7r7"] Oct 10 15:11:24 crc kubenswrapper[4788]: I1010 15:11:24.971178 4788 scope.go:117] "RemoveContainer" containerID="42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0" Oct 10 15:11:25 crc kubenswrapper[4788]: I1010 15:11:25.009618 4788 scope.go:117] "RemoveContainer" containerID="42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882" Oct 10 15:11:25 crc kubenswrapper[4788]: E1010 15:11:25.010087 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882\": container with ID starting with 42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882 not found: ID does not exist" containerID="42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882" Oct 10 15:11:25 crc kubenswrapper[4788]: I1010 15:11:25.010133 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882"} err="failed to get container status \"42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882\": rpc error: code = NotFound desc = could not find container \"42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882\": container with ID starting with 42c2a043be70a38a647708c7657be046f9891b29ef5d603857fcc22d875cd882 not found: ID does not exist" Oct 10 15:11:25 crc kubenswrapper[4788]: I1010 15:11:25.010196 4788 scope.go:117] "RemoveContainer" containerID="2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6" Oct 10 15:11:25 crc kubenswrapper[4788]: E1010 15:11:25.010589 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6\": container with ID starting with 2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6 not found: ID does not exist" containerID="2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6" Oct 10 15:11:25 crc kubenswrapper[4788]: I1010 15:11:25.010620 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6"} err="failed to get container status \"2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6\": rpc error: code = NotFound desc = could not find container \"2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6\": container with ID starting with 2c24d9bdf7e7e2a35bef69283111aa11c8f48091347e115c3cd6b49795c600d6 not found: ID does not exist" Oct 10 15:11:25 crc kubenswrapper[4788]: I1010 15:11:25.010640 4788 scope.go:117] "RemoveContainer" containerID="42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0" Oct 10 15:11:25 crc kubenswrapper[4788]: E1010 15:11:25.010905 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0\": container with ID starting with 42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0 not found: ID does not exist" containerID="42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0" Oct 10 15:11:25 crc kubenswrapper[4788]: I1010 15:11:25.010935 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0"} err="failed to get container status \"42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0\": rpc error: code = NotFound desc = could not find container \"42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0\": container with ID starting with 42d9bf7913e2be4f917238fb4da6817c84f4479a07eef6966f61a51826de83d0 not found: ID does not exist" Oct 10 15:11:26 crc kubenswrapper[4788]: I1010 15:11:26.249432 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" path="/var/lib/kubelet/pods/ff7d78a9-cb91-47e4-93bb-19d1b112a329/volumes" Oct 10 15:11:28 crc kubenswrapper[4788]: I1010 15:11:28.237107 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:11:28 crc kubenswrapper[4788]: E1010 15:11:28.237515 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:11:42 crc kubenswrapper[4788]: I1010 15:11:42.234271 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:11:42 crc kubenswrapper[4788]: E1010 15:11:42.235005 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:11:56 crc kubenswrapper[4788]: I1010 15:11:56.234875 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:11:56 crc kubenswrapper[4788]: E1010 15:11:56.235818 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.211560 4788 scope.go:117] "RemoveContainer" containerID="e71c6da1b980b5831fd3e520526064c0e485bbe80ec60e1dc2ac9a1be89dab0e" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.236687 4788 scope.go:117] "RemoveContainer" containerID="8830b296e336ae0bb205059ae2ae08d16b5ce3f0c6cc07992145bafa1eddc0ce" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.288893 4788 scope.go:117] "RemoveContainer" containerID="095bf53a298c6f70b38e43a9a5978cec79e29c28c3436fdae08686599935c768" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.352997 4788 scope.go:117] "RemoveContainer" containerID="e09f1669e3a187300831d30e49a102971bd1cc7640e3796649e5f7fdfbec27e0" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.376065 4788 scope.go:117] "RemoveContainer" containerID="59d414f971d9049d275be0eca426c5c85df49ffbd6a3195e6c288e33645927bc" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.423606 4788 scope.go:117] "RemoveContainer" containerID="7924633e2252255a2f8fe0feed662aecd204a9fa8c9987dfa08717028cec687a" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.451275 4788 scope.go:117] "RemoveContainer" containerID="27deb31db11e825d80667118e5461f9ff9e69e6d1632a0d32d3510499e4941b1" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.472090 4788 scope.go:117] "RemoveContainer" containerID="7ecafbd5dc24dad827ac28f9bc73bf2383bcffe7b45d5665ac712e4299074544" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.492241 4788 scope.go:117] "RemoveContainer" containerID="f1acd019454438f9f318884feab766dfda302a7f2270489b71e0fc3660adf204" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.509979 4788 scope.go:117] "RemoveContainer" containerID="bfc0913bcd1ab1498eb49921a29998fe79941d93b829b0a8fb7368a4fa4e2fdd" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.544502 4788 scope.go:117] "RemoveContainer" containerID="e39944b79b0c075079e9a288c1893a4cdbd681cc744af6cff568e6a8933d1c22" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.566411 4788 scope.go:117] "RemoveContainer" containerID="f98704522449389d2af1c513f520e6f82ebedfbe293c1adf7038d2875138c32a" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.582013 4788 scope.go:117] "RemoveContainer" containerID="cd905b3981551c017fee3cce5d78db27acb3a006b0a289d22d1980255e28acb8" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.599632 4788 scope.go:117] "RemoveContainer" containerID="a39f5358541468b172ce1b2d14be325d374226c5cb18b04385a6e40974271560" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.619122 4788 scope.go:117] "RemoveContainer" containerID="441c3dc286f891485385a909d54daf00c136c609bb2e0e40814c55b244268d27" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.633267 4788 scope.go:117] "RemoveContainer" containerID="907c6c53800cba4ebb91526f1fea5664b224ece7220cf782500a49298e929afc" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.657101 4788 scope.go:117] "RemoveContainer" containerID="183b4d966d60fe3ff4c7e99782bb6afa87a35b173ac80cac610a4890b558dd1c" Oct 10 15:12:08 crc kubenswrapper[4788]: I1010 15:12:08.679549 4788 scope.go:117] "RemoveContainer" containerID="8e6c023e337efd4e7921b84ae53a248be93d4aba90a3ae0477341ae598b257a9" Oct 10 15:12:11 crc kubenswrapper[4788]: I1010 15:12:11.234721 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:12:11 crc kubenswrapper[4788]: E1010 15:12:11.235775 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:12:25 crc kubenswrapper[4788]: I1010 15:12:25.235046 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:12:25 crc kubenswrapper[4788]: E1010 15:12:25.235620 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:12:36 crc kubenswrapper[4788]: I1010 15:12:36.234079 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:12:36 crc kubenswrapper[4788]: E1010 15:12:36.234995 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:12:51 crc kubenswrapper[4788]: I1010 15:12:51.234232 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:12:51 crc kubenswrapper[4788]: E1010 15:12:51.235272 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:13:04 crc kubenswrapper[4788]: I1010 15:13:04.233988 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:13:04 crc kubenswrapper[4788]: E1010 15:13:04.234949 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:13:08 crc kubenswrapper[4788]: I1010 15:13:08.928387 4788 scope.go:117] "RemoveContainer" containerID="99ac1c7166ca7cb51003167ee8e34ec922f167234c00c272405405fcad496e3a" Oct 10 15:13:08 crc kubenswrapper[4788]: I1010 15:13:08.964064 4788 scope.go:117] "RemoveContainer" containerID="75754fabec2a957f5473c8b6d0e269b61de4bd802556fd25db9f3d0c62cdf27f" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.003496 4788 scope.go:117] "RemoveContainer" containerID="ec39cee08096b21f5fb188e2f4be5f41b034a3a3eb4ccdb2d9860218975f6b3f" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.033509 4788 scope.go:117] "RemoveContainer" containerID="c74559860631b8471c48eefeb8cff5db5a27cbd9c356e3b6d62ee684f03c0f98" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.077491 4788 scope.go:117] "RemoveContainer" containerID="dd729b335f5887ad2a47d9138d9341e607bccc05df0bd17180c3c60156b9f713" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.113188 4788 scope.go:117] "RemoveContainer" containerID="a2c0776de729185f9cd35be86586a5c1c49c778ed6a5e56f595dcf61caf342d5" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.144904 4788 scope.go:117] "RemoveContainer" containerID="716fea2b9b562081878c7543dd14ca1d4276b93e4640be2eb106923b236a4f59" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.165699 4788 scope.go:117] "RemoveContainer" containerID="7e46b1cf1f8c6202fa33e2921e3f1273618b996ccc581604d916fffc10f5504e" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.221415 4788 scope.go:117] "RemoveContainer" containerID="5c1c142fd945aca3169ac20af4021c94a352596999aa62187f338f4da44d457b" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.252387 4788 scope.go:117] "RemoveContainer" containerID="6129639cc6300aaa58f2036c176db47627bdf756435da9fb378d8d6f03f3906f" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.281350 4788 scope.go:117] "RemoveContainer" containerID="738718baa0a2bcb4b08f63698f8aa53d2a8d14acac079c8121fb629288d77d15" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.309261 4788 scope.go:117] "RemoveContainer" containerID="a2a9700b17de8390730939a608c45fe4b3b60e9b1e67395aceadf7ac2576f14f" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.328871 4788 scope.go:117] "RemoveContainer" containerID="af0d26bb8887c19c55d06a25ecfaf257af7796e1125a3c42fd81381267d630df" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.361382 4788 scope.go:117] "RemoveContainer" containerID="1e94fd37cb99c18e9e70ea41a5932cd392fae9e72bcd73f02398b3f4d1d01fca" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.413051 4788 scope.go:117] "RemoveContainer" containerID="10a6407ac6ed46784d87d17ac1f81a2239476dc6186c82ebe7235c2468e4f18c" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.429895 4788 scope.go:117] "RemoveContainer" containerID="3f97b52d47c54bb90b4b7160365c3a49e2c886ca8a65ee3949f187e17ec1123b" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.444850 4788 scope.go:117] "RemoveContainer" containerID="677ff4519271b2895389af925e7a410ae00b4a773ad88355cb333084a7d66722" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.460027 4788 scope.go:117] "RemoveContainer" containerID="75bc72d77ca51763e142eda1335a09a768c99c2bf9ade041fdec8f17c2aa8354" Oct 10 15:13:09 crc kubenswrapper[4788]: I1010 15:13:09.480822 4788 scope.go:117] "RemoveContainer" containerID="824819b6460de36355916687ebd695df87d0acd9246071dee316aeadb206bf16" Oct 10 15:13:18 crc kubenswrapper[4788]: I1010 15:13:18.242627 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:13:18 crc kubenswrapper[4788]: E1010 15:13:18.243949 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:13:32 crc kubenswrapper[4788]: I1010 15:13:32.233425 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:13:32 crc kubenswrapper[4788]: E1010 15:13:32.234087 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:13:46 crc kubenswrapper[4788]: I1010 15:13:46.234621 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:13:46 crc kubenswrapper[4788]: E1010 15:13:46.235445 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:13:57 crc kubenswrapper[4788]: I1010 15:13:57.235117 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:13:57 crc kubenswrapper[4788]: E1010 15:13:57.236239 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:14:08 crc kubenswrapper[4788]: I1010 15:14:08.238275 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:14:08 crc kubenswrapper[4788]: E1010 15:14:08.239243 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:14:09 crc kubenswrapper[4788]: I1010 15:14:09.737356 4788 scope.go:117] "RemoveContainer" containerID="155460bd9f7f7a8b06685eb2550e5f97ae4172a9cd51c3ea8ed6a80fa5b5287e" Oct 10 15:14:09 crc kubenswrapper[4788]: I1010 15:14:09.771202 4788 scope.go:117] "RemoveContainer" containerID="1dec38763ab92d684d95ae78a000552b0eab6a7c2f48c775ab127484d305983d" Oct 10 15:14:09 crc kubenswrapper[4788]: I1010 15:14:09.826627 4788 scope.go:117] "RemoveContainer" containerID="51707ce7890bf73c148c3998b37c2bd349e7afe315f7fbe5508284b03085928d" Oct 10 15:14:09 crc kubenswrapper[4788]: I1010 15:14:09.877360 4788 scope.go:117] "RemoveContainer" containerID="4104f4cfee1edb339a3c5824fe82691a528e1cc74ee8a2ec60a5c133c43e91ee" Oct 10 15:14:09 crc kubenswrapper[4788]: I1010 15:14:09.898490 4788 scope.go:117] "RemoveContainer" containerID="3e0368d97024ca12075d24fda31d7d4eead57732acfefdcbd6d20c33d4f7f6f7" Oct 10 15:14:09 crc kubenswrapper[4788]: I1010 15:14:09.915552 4788 scope.go:117] "RemoveContainer" containerID="e0ce65ec0009a40c1eb733283f7b65324c71d464e88882462d80a65bd07ae1af" Oct 10 15:14:09 crc kubenswrapper[4788]: I1010 15:14:09.935285 4788 scope.go:117] "RemoveContainer" containerID="9c4939632ebd89196bb2dfa61693556ec6deaee5d6432afbfdb9352a9b188e2a" Oct 10 15:14:19 crc kubenswrapper[4788]: I1010 15:14:19.234051 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:14:19 crc kubenswrapper[4788]: E1010 15:14:19.235420 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:14:33 crc kubenswrapper[4788]: I1010 15:14:33.234660 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:14:33 crc kubenswrapper[4788]: I1010 15:14:33.870416 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"2db1babb5b22825b16cd8f5b57897f948f0ae665e548e72a03b5b815c79d0f5d"} Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.148411 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s"] Oct 10 15:15:00 crc kubenswrapper[4788]: E1010 15:15:00.149205 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerName="extract-content" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.149219 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerName="extract-content" Oct 10 15:15:00 crc kubenswrapper[4788]: E1010 15:15:00.149249 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerName="extract-utilities" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.149256 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerName="extract-utilities" Oct 10 15:15:00 crc kubenswrapper[4788]: E1010 15:15:00.149266 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerName="registry-server" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.149273 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerName="registry-server" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.149421 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7d78a9-cb91-47e4-93bb-19d1b112a329" containerName="registry-server" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.150326 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.153264 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.153544 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.164753 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s"] Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.204537 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hbls\" (UniqueName: \"kubernetes.io/projected/1f5943a2-fc93-4c06-bd9e-bddfafa70008-kube-api-access-8hbls\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.204575 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f5943a2-fc93-4c06-bd9e-bddfafa70008-config-volume\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.204627 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f5943a2-fc93-4c06-bd9e-bddfafa70008-secret-volume\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.305581 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hbls\" (UniqueName: \"kubernetes.io/projected/1f5943a2-fc93-4c06-bd9e-bddfafa70008-kube-api-access-8hbls\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.305959 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f5943a2-fc93-4c06-bd9e-bddfafa70008-config-volume\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.306043 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f5943a2-fc93-4c06-bd9e-bddfafa70008-secret-volume\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.306914 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f5943a2-fc93-4c06-bd9e-bddfafa70008-config-volume\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.312841 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f5943a2-fc93-4c06-bd9e-bddfafa70008-secret-volume\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.321196 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hbls\" (UniqueName: \"kubernetes.io/projected/1f5943a2-fc93-4c06-bd9e-bddfafa70008-kube-api-access-8hbls\") pod \"collect-profiles-29335155-kqc4s\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.543246 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:00 crc kubenswrapper[4788]: I1010 15:15:00.939288 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s"] Oct 10 15:15:01 crc kubenswrapper[4788]: I1010 15:15:01.131468 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" event={"ID":"1f5943a2-fc93-4c06-bd9e-bddfafa70008","Type":"ContainerStarted","Data":"ab1d3fe6a8c266881d42fd2b1a9fbe4f5112d4b810842047600cbe7d2a71dffa"} Oct 10 15:15:01 crc kubenswrapper[4788]: I1010 15:15:01.131525 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" event={"ID":"1f5943a2-fc93-4c06-bd9e-bddfafa70008","Type":"ContainerStarted","Data":"164cc371d15fdf304ee91c973fe4e1a11db06fd015ebba6d216cdb90a82f5f8a"} Oct 10 15:15:01 crc kubenswrapper[4788]: I1010 15:15:01.152045 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" podStartSLOduration=1.152021875 podStartE2EDuration="1.152021875s" podCreationTimestamp="2025-10-10 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 15:15:01.14923403 +0000 UTC m=+1803.598949578" watchObservedRunningTime="2025-10-10 15:15:01.152021875 +0000 UTC m=+1803.601737433" Oct 10 15:15:02 crc kubenswrapper[4788]: I1010 15:15:02.139355 4788 generic.go:334] "Generic (PLEG): container finished" podID="1f5943a2-fc93-4c06-bd9e-bddfafa70008" containerID="ab1d3fe6a8c266881d42fd2b1a9fbe4f5112d4b810842047600cbe7d2a71dffa" exitCode=0 Oct 10 15:15:02 crc kubenswrapper[4788]: I1010 15:15:02.139424 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" event={"ID":"1f5943a2-fc93-4c06-bd9e-bddfafa70008","Type":"ContainerDied","Data":"ab1d3fe6a8c266881d42fd2b1a9fbe4f5112d4b810842047600cbe7d2a71dffa"} Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.432546 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.448610 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hbls\" (UniqueName: \"kubernetes.io/projected/1f5943a2-fc93-4c06-bd9e-bddfafa70008-kube-api-access-8hbls\") pod \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.450468 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f5943a2-fc93-4c06-bd9e-bddfafa70008-config-volume" (OuterVolumeSpecName: "config-volume") pod "1f5943a2-fc93-4c06-bd9e-bddfafa70008" (UID: "1f5943a2-fc93-4c06-bd9e-bddfafa70008"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.451298 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f5943a2-fc93-4c06-bd9e-bddfafa70008-config-volume\") pod \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.451391 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f5943a2-fc93-4c06-bd9e-bddfafa70008-secret-volume\") pod \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\" (UID: \"1f5943a2-fc93-4c06-bd9e-bddfafa70008\") " Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.454750 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f5943a2-fc93-4c06-bd9e-bddfafa70008-kube-api-access-8hbls" (OuterVolumeSpecName: "kube-api-access-8hbls") pod "1f5943a2-fc93-4c06-bd9e-bddfafa70008" (UID: "1f5943a2-fc93-4c06-bd9e-bddfafa70008"). InnerVolumeSpecName "kube-api-access-8hbls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.457297 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f5943a2-fc93-4c06-bd9e-bddfafa70008-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1f5943a2-fc93-4c06-bd9e-bddfafa70008" (UID: "1f5943a2-fc93-4c06-bd9e-bddfafa70008"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.553672 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hbls\" (UniqueName: \"kubernetes.io/projected/1f5943a2-fc93-4c06-bd9e-bddfafa70008-kube-api-access-8hbls\") on node \"crc\" DevicePath \"\"" Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.553706 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f5943a2-fc93-4c06-bd9e-bddfafa70008-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 15:15:03 crc kubenswrapper[4788]: I1010 15:15:03.553718 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f5943a2-fc93-4c06-bd9e-bddfafa70008-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 15:15:04 crc kubenswrapper[4788]: I1010 15:15:04.158606 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" event={"ID":"1f5943a2-fc93-4c06-bd9e-bddfafa70008","Type":"ContainerDied","Data":"164cc371d15fdf304ee91c973fe4e1a11db06fd015ebba6d216cdb90a82f5f8a"} Oct 10 15:15:04 crc kubenswrapper[4788]: I1010 15:15:04.158651 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="164cc371d15fdf304ee91c973fe4e1a11db06fd015ebba6d216cdb90a82f5f8a" Oct 10 15:15:04 crc kubenswrapper[4788]: I1010 15:15:04.158715 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.102329 4788 scope.go:117] "RemoveContainer" containerID="5bc7218d76c149b7925383f9d5d6ab8180d2bb5b1b399cf25e15bef72a048be8" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.117997 4788 scope.go:117] "RemoveContainer" containerID="598ec17f8706ac8a7613481f8dfaf31904983cb92d6010eefbf1ac47efcbb014" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.134110 4788 scope.go:117] "RemoveContainer" containerID="79447d873480f8db4080543745a5c249b2cbec57b2c886298cf38288010480c5" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.150559 4788 scope.go:117] "RemoveContainer" containerID="fc75d105447e3e2c43eec1601b71054b73baff8f958dbf327e3eb0680d5cdba4" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.177918 4788 scope.go:117] "RemoveContainer" containerID="b2c6bf83b23d7eb9e36e2d8a2de6b0fcfeb53ed06aa0aaeb6182824c8e84405f" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.237915 4788 scope.go:117] "RemoveContainer" containerID="edf59e020f4a2e0970646b96aab104bff00861373f9c92c494f167dd813d4c13" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.264433 4788 scope.go:117] "RemoveContainer" containerID="6065428399b804c4b4b31556aa5431d788e52a568e70173b39bb05a529f66b04" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.286351 4788 scope.go:117] "RemoveContainer" containerID="fab44d16043a2503f608627193df1f5950c0f9017dc038519c69453dcac75e2f" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.308509 4788 scope.go:117] "RemoveContainer" containerID="0a1d2c1f952efc125bdddea9f7888d1cfbf3c24aa07f6a81eb81823168495ca6" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.326762 4788 scope.go:117] "RemoveContainer" containerID="ae99f44876fccdb0bd4b1bbbc88a3459a502c4174dd932d08139e04e243eaf54" Oct 10 15:15:10 crc kubenswrapper[4788]: I1010 15:15:10.350442 4788 scope.go:117] "RemoveContainer" containerID="cc438a9247888adf5836596dc8ce1d20b78dd2a805885fd538540aba1444f53e" Oct 10 15:16:59 crc kubenswrapper[4788]: I1010 15:16:59.406495 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:16:59 crc kubenswrapper[4788]: I1010 15:16:59.407019 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:17:29 crc kubenswrapper[4788]: I1010 15:17:29.405989 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:17:29 crc kubenswrapper[4788]: I1010 15:17:29.406543 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:17:59 crc kubenswrapper[4788]: I1010 15:17:59.406529 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:17:59 crc kubenswrapper[4788]: I1010 15:17:59.409531 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:17:59 crc kubenswrapper[4788]: I1010 15:17:59.409801 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:17:59 crc kubenswrapper[4788]: I1010 15:17:59.411231 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2db1babb5b22825b16cd8f5b57897f948f0ae665e548e72a03b5b815c79d0f5d"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:17:59 crc kubenswrapper[4788]: I1010 15:17:59.411769 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://2db1babb5b22825b16cd8f5b57897f948f0ae665e548e72a03b5b815c79d0f5d" gracePeriod=600 Oct 10 15:17:59 crc kubenswrapper[4788]: I1010 15:17:59.706759 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="2db1babb5b22825b16cd8f5b57897f948f0ae665e548e72a03b5b815c79d0f5d" exitCode=0 Oct 10 15:17:59 crc kubenswrapper[4788]: I1010 15:17:59.706936 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"2db1babb5b22825b16cd8f5b57897f948f0ae665e548e72a03b5b815c79d0f5d"} Oct 10 15:17:59 crc kubenswrapper[4788]: I1010 15:17:59.707055 4788 scope.go:117] "RemoveContainer" containerID="2cb5f6f42cf805ce3cd480726b9a159ccfeba06715d23e8b76a279e4dba84782" Oct 10 15:18:00 crc kubenswrapper[4788]: I1010 15:18:00.736613 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49"} Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.626440 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x8gpz"] Oct 10 15:19:55 crc kubenswrapper[4788]: E1010 15:19:55.628497 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f5943a2-fc93-4c06-bd9e-bddfafa70008" containerName="collect-profiles" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.628529 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f5943a2-fc93-4c06-bd9e-bddfafa70008" containerName="collect-profiles" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.628872 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f5943a2-fc93-4c06-bd9e-bddfafa70008" containerName="collect-profiles" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.631030 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.667947 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8gpz"] Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.750480 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc54f\" (UniqueName: \"kubernetes.io/projected/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-kube-api-access-lc54f\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.750583 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-catalog-content\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.750678 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-utilities\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.853303 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc54f\" (UniqueName: \"kubernetes.io/projected/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-kube-api-access-lc54f\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.853377 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-catalog-content\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.853437 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-utilities\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.854046 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-utilities\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.854472 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-catalog-content\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.885206 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc54f\" (UniqueName: \"kubernetes.io/projected/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-kube-api-access-lc54f\") pod \"redhat-operators-x8gpz\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:55 crc kubenswrapper[4788]: I1010 15:19:55.988060 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:19:56 crc kubenswrapper[4788]: I1010 15:19:56.517577 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8gpz"] Oct 10 15:19:56 crc kubenswrapper[4788]: I1010 15:19:56.775885 4788 generic.go:334] "Generic (PLEG): container finished" podID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerID="7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c" exitCode=0 Oct 10 15:19:56 crc kubenswrapper[4788]: I1010 15:19:56.775946 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8gpz" event={"ID":"9b47ebd7-75a5-4eaa-9abe-e439196bbeda","Type":"ContainerDied","Data":"7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c"} Oct 10 15:19:56 crc kubenswrapper[4788]: I1010 15:19:56.775982 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8gpz" event={"ID":"9b47ebd7-75a5-4eaa-9abe-e439196bbeda","Type":"ContainerStarted","Data":"e44e4e552fe850f0014cab01cf526c2300f19eedee1189cd673dd3e14d5f7802"} Oct 10 15:19:56 crc kubenswrapper[4788]: I1010 15:19:56.778628 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 15:19:57 crc kubenswrapper[4788]: I1010 15:19:57.789077 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8gpz" event={"ID":"9b47ebd7-75a5-4eaa-9abe-e439196bbeda","Type":"ContainerStarted","Data":"9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee"} Oct 10 15:19:58 crc kubenswrapper[4788]: I1010 15:19:58.806007 4788 generic.go:334] "Generic (PLEG): container finished" podID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerID="9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee" exitCode=0 Oct 10 15:19:58 crc kubenswrapper[4788]: I1010 15:19:58.806104 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8gpz" event={"ID":"9b47ebd7-75a5-4eaa-9abe-e439196bbeda","Type":"ContainerDied","Data":"9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee"} Oct 10 15:19:59 crc kubenswrapper[4788]: I1010 15:19:59.406132 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:19:59 crc kubenswrapper[4788]: I1010 15:19:59.406888 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:19:59 crc kubenswrapper[4788]: I1010 15:19:59.822460 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8gpz" event={"ID":"9b47ebd7-75a5-4eaa-9abe-e439196bbeda","Type":"ContainerStarted","Data":"8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea"} Oct 10 15:19:59 crc kubenswrapper[4788]: I1010 15:19:59.852054 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x8gpz" podStartSLOduration=2.329278703 podStartE2EDuration="4.852028898s" podCreationTimestamp="2025-10-10 15:19:55 +0000 UTC" firstStartedPulling="2025-10-10 15:19:56.778376963 +0000 UTC m=+2099.228092501" lastFinishedPulling="2025-10-10 15:19:59.301127108 +0000 UTC m=+2101.750842696" observedRunningTime="2025-10-10 15:19:59.850724963 +0000 UTC m=+2102.300440521" watchObservedRunningTime="2025-10-10 15:19:59.852028898 +0000 UTC m=+2102.301744496" Oct 10 15:20:05 crc kubenswrapper[4788]: I1010 15:20:05.988459 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:20:05 crc kubenswrapper[4788]: I1010 15:20:05.989399 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:20:06 crc kubenswrapper[4788]: I1010 15:20:06.046334 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:20:06 crc kubenswrapper[4788]: I1010 15:20:06.957111 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:20:07 crc kubenswrapper[4788]: I1010 15:20:07.019478 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x8gpz"] Oct 10 15:20:08 crc kubenswrapper[4788]: I1010 15:20:08.897695 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x8gpz" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerName="registry-server" containerID="cri-o://8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea" gracePeriod=2 Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.484414 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.681716 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-utilities\") pod \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.681798 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-catalog-content\") pod \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.682012 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc54f\" (UniqueName: \"kubernetes.io/projected/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-kube-api-access-lc54f\") pod \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\" (UID: \"9b47ebd7-75a5-4eaa-9abe-e439196bbeda\") " Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.684734 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-utilities" (OuterVolumeSpecName: "utilities") pod "9b47ebd7-75a5-4eaa-9abe-e439196bbeda" (UID: "9b47ebd7-75a5-4eaa-9abe-e439196bbeda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.690402 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-kube-api-access-lc54f" (OuterVolumeSpecName: "kube-api-access-lc54f") pod "9b47ebd7-75a5-4eaa-9abe-e439196bbeda" (UID: "9b47ebd7-75a5-4eaa-9abe-e439196bbeda"). InnerVolumeSpecName "kube-api-access-lc54f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.786684 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc54f\" (UniqueName: \"kubernetes.io/projected/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-kube-api-access-lc54f\") on node \"crc\" DevicePath \"\"" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.786733 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.828121 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b47ebd7-75a5-4eaa-9abe-e439196bbeda" (UID: "9b47ebd7-75a5-4eaa-9abe-e439196bbeda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.888363 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b47ebd7-75a5-4eaa-9abe-e439196bbeda-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.913742 4788 generic.go:334] "Generic (PLEG): container finished" podID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerID="8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea" exitCode=0 Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.913819 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8gpz" event={"ID":"9b47ebd7-75a5-4eaa-9abe-e439196bbeda","Type":"ContainerDied","Data":"8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea"} Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.913870 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8gpz" event={"ID":"9b47ebd7-75a5-4eaa-9abe-e439196bbeda","Type":"ContainerDied","Data":"e44e4e552fe850f0014cab01cf526c2300f19eedee1189cd673dd3e14d5f7802"} Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.913905 4788 scope.go:117] "RemoveContainer" containerID="8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.913933 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8gpz" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.948749 4788 scope.go:117] "RemoveContainer" containerID="9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.987725 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x8gpz"] Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.992800 4788 scope.go:117] "RemoveContainer" containerID="7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c" Oct 10 15:20:09 crc kubenswrapper[4788]: I1010 15:20:09.996490 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x8gpz"] Oct 10 15:20:10 crc kubenswrapper[4788]: I1010 15:20:10.023689 4788 scope.go:117] "RemoveContainer" containerID="8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea" Oct 10 15:20:10 crc kubenswrapper[4788]: E1010 15:20:10.024581 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea\": container with ID starting with 8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea not found: ID does not exist" containerID="8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea" Oct 10 15:20:10 crc kubenswrapper[4788]: I1010 15:20:10.024649 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea"} err="failed to get container status \"8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea\": rpc error: code = NotFound desc = could not find container \"8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea\": container with ID starting with 8946e060ba7b793c5f1651bbf3fbb1af055d1448c369bfed71ba9ef66bc6e3ea not found: ID does not exist" Oct 10 15:20:10 crc kubenswrapper[4788]: I1010 15:20:10.024692 4788 scope.go:117] "RemoveContainer" containerID="9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee" Oct 10 15:20:10 crc kubenswrapper[4788]: E1010 15:20:10.025493 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee\": container with ID starting with 9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee not found: ID does not exist" containerID="9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee" Oct 10 15:20:10 crc kubenswrapper[4788]: I1010 15:20:10.025548 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee"} err="failed to get container status \"9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee\": rpc error: code = NotFound desc = could not find container \"9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee\": container with ID starting with 9dadd2a7d812881f8430bf0ba3f5de9a032f9ed170a27c3a5b87f9a33bf3cdee not found: ID does not exist" Oct 10 15:20:10 crc kubenswrapper[4788]: I1010 15:20:10.025585 4788 scope.go:117] "RemoveContainer" containerID="7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c" Oct 10 15:20:10 crc kubenswrapper[4788]: E1010 15:20:10.026060 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c\": container with ID starting with 7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c not found: ID does not exist" containerID="7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c" Oct 10 15:20:10 crc kubenswrapper[4788]: I1010 15:20:10.026111 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c"} err="failed to get container status \"7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c\": rpc error: code = NotFound desc = could not find container \"7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c\": container with ID starting with 7d7ccc480ba972aa94d0efab1b6ac4d41269052200d60c998ea0675b9fa4a94c not found: ID does not exist" Oct 10 15:20:10 crc kubenswrapper[4788]: I1010 15:20:10.245310 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" path="/var/lib/kubelet/pods/9b47ebd7-75a5-4eaa-9abe-e439196bbeda/volumes" Oct 10 15:20:29 crc kubenswrapper[4788]: I1010 15:20:29.406822 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:20:29 crc kubenswrapper[4788]: I1010 15:20:29.407478 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.583584 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d47q5"] Oct 10 15:20:31 crc kubenswrapper[4788]: E1010 15:20:31.584985 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerName="registry-server" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.585026 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerName="registry-server" Oct 10 15:20:31 crc kubenswrapper[4788]: E1010 15:20:31.585060 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerName="extract-utilities" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.585079 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerName="extract-utilities" Oct 10 15:20:31 crc kubenswrapper[4788]: E1010 15:20:31.585179 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerName="extract-content" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.585201 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerName="extract-content" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.585651 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b47ebd7-75a5-4eaa-9abe-e439196bbeda" containerName="registry-server" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.588375 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.598589 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d47q5"] Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.673087 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-utilities\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.673449 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-catalog-content\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.673533 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw2lt\" (UniqueName: \"kubernetes.io/projected/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-kube-api-access-hw2lt\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.775452 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-catalog-content\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.775529 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw2lt\" (UniqueName: \"kubernetes.io/projected/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-kube-api-access-hw2lt\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.775594 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-utilities\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.776101 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-catalog-content\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.776191 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-utilities\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.800809 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw2lt\" (UniqueName: \"kubernetes.io/projected/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-kube-api-access-hw2lt\") pod \"certified-operators-d47q5\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:31 crc kubenswrapper[4788]: I1010 15:20:31.934770 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:32 crc kubenswrapper[4788]: I1010 15:20:32.229919 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d47q5"] Oct 10 15:20:33 crc kubenswrapper[4788]: I1010 15:20:33.142564 4788 generic.go:334] "Generic (PLEG): container finished" podID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerID="ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8" exitCode=0 Oct 10 15:20:33 crc kubenswrapper[4788]: I1010 15:20:33.142789 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d47q5" event={"ID":"fa1f8039-d1cb-466d-9dc2-8881eeeefa36","Type":"ContainerDied","Data":"ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8"} Oct 10 15:20:33 crc kubenswrapper[4788]: I1010 15:20:33.143075 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d47q5" event={"ID":"fa1f8039-d1cb-466d-9dc2-8881eeeefa36","Type":"ContainerStarted","Data":"53dff14f9b4be2379e4bfedead629e939acd54af816cdddfaab3d6f7d0c05265"} Oct 10 15:20:34 crc kubenswrapper[4788]: I1010 15:20:34.158252 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d47q5" event={"ID":"fa1f8039-d1cb-466d-9dc2-8881eeeefa36","Type":"ContainerStarted","Data":"c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9"} Oct 10 15:20:35 crc kubenswrapper[4788]: I1010 15:20:35.169388 4788 generic.go:334] "Generic (PLEG): container finished" podID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerID="c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9" exitCode=0 Oct 10 15:20:35 crc kubenswrapper[4788]: I1010 15:20:35.169463 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d47q5" event={"ID":"fa1f8039-d1cb-466d-9dc2-8881eeeefa36","Type":"ContainerDied","Data":"c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9"} Oct 10 15:20:36 crc kubenswrapper[4788]: I1010 15:20:36.183694 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d47q5" event={"ID":"fa1f8039-d1cb-466d-9dc2-8881eeeefa36","Type":"ContainerStarted","Data":"5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd"} Oct 10 15:20:36 crc kubenswrapper[4788]: I1010 15:20:36.215387 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d47q5" podStartSLOduration=2.719225389 podStartE2EDuration="5.21536627s" podCreationTimestamp="2025-10-10 15:20:31 +0000 UTC" firstStartedPulling="2025-10-10 15:20:33.145738554 +0000 UTC m=+2135.595454132" lastFinishedPulling="2025-10-10 15:20:35.641879445 +0000 UTC m=+2138.091595013" observedRunningTime="2025-10-10 15:20:36.213683345 +0000 UTC m=+2138.663398933" watchObservedRunningTime="2025-10-10 15:20:36.21536627 +0000 UTC m=+2138.665081828" Oct 10 15:20:41 crc kubenswrapper[4788]: I1010 15:20:41.935244 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:41 crc kubenswrapper[4788]: I1010 15:20:41.939252 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:42 crc kubenswrapper[4788]: I1010 15:20:42.001296 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:42 crc kubenswrapper[4788]: I1010 15:20:42.306494 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:42 crc kubenswrapper[4788]: I1010 15:20:42.372735 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d47q5"] Oct 10 15:20:44 crc kubenswrapper[4788]: I1010 15:20:44.265226 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d47q5" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerName="registry-server" containerID="cri-o://5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd" gracePeriod=2 Oct 10 15:20:44 crc kubenswrapper[4788]: I1010 15:20:44.823258 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:44 crc kubenswrapper[4788]: I1010 15:20:44.928002 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-utilities\") pod \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " Oct 10 15:20:44 crc kubenswrapper[4788]: I1010 15:20:44.928188 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-catalog-content\") pod \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " Oct 10 15:20:44 crc kubenswrapper[4788]: I1010 15:20:44.928223 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw2lt\" (UniqueName: \"kubernetes.io/projected/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-kube-api-access-hw2lt\") pod \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\" (UID: \"fa1f8039-d1cb-466d-9dc2-8881eeeefa36\") " Oct 10 15:20:44 crc kubenswrapper[4788]: I1010 15:20:44.929405 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-utilities" (OuterVolumeSpecName: "utilities") pod "fa1f8039-d1cb-466d-9dc2-8881eeeefa36" (UID: "fa1f8039-d1cb-466d-9dc2-8881eeeefa36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:20:44 crc kubenswrapper[4788]: I1010 15:20:44.938450 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-kube-api-access-hw2lt" (OuterVolumeSpecName: "kube-api-access-hw2lt") pod "fa1f8039-d1cb-466d-9dc2-8881eeeefa36" (UID: "fa1f8039-d1cb-466d-9dc2-8881eeeefa36"). InnerVolumeSpecName "kube-api-access-hw2lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.005044 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa1f8039-d1cb-466d-9dc2-8881eeeefa36" (UID: "fa1f8039-d1cb-466d-9dc2-8881eeeefa36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.029858 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.029914 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.029938 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw2lt\" (UniqueName: \"kubernetes.io/projected/fa1f8039-d1cb-466d-9dc2-8881eeeefa36-kube-api-access-hw2lt\") on node \"crc\" DevicePath \"\"" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.278007 4788 generic.go:334] "Generic (PLEG): container finished" podID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerID="5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd" exitCode=0 Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.278059 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d47q5" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.278071 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d47q5" event={"ID":"fa1f8039-d1cb-466d-9dc2-8881eeeefa36","Type":"ContainerDied","Data":"5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd"} Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.278119 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d47q5" event={"ID":"fa1f8039-d1cb-466d-9dc2-8881eeeefa36","Type":"ContainerDied","Data":"53dff14f9b4be2379e4bfedead629e939acd54af816cdddfaab3d6f7d0c05265"} Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.278267 4788 scope.go:117] "RemoveContainer" containerID="5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.307539 4788 scope.go:117] "RemoveContainer" containerID="c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.314909 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d47q5"] Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.323288 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d47q5"] Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.343641 4788 scope.go:117] "RemoveContainer" containerID="ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.369563 4788 scope.go:117] "RemoveContainer" containerID="5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd" Oct 10 15:20:45 crc kubenswrapper[4788]: E1010 15:20:45.370103 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd\": container with ID starting with 5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd not found: ID does not exist" containerID="5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.370380 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd"} err="failed to get container status \"5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd\": rpc error: code = NotFound desc = could not find container \"5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd\": container with ID starting with 5c318ed36f653b1f0650c4eb48b34cdc53ea4f4bb07d510843ecfd80a9a6fffd not found: ID does not exist" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.370430 4788 scope.go:117] "RemoveContainer" containerID="c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9" Oct 10 15:20:45 crc kubenswrapper[4788]: E1010 15:20:45.371081 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9\": container with ID starting with c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9 not found: ID does not exist" containerID="c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.371220 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9"} err="failed to get container status \"c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9\": rpc error: code = NotFound desc = could not find container \"c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9\": container with ID starting with c3487da154c4587d4ce76c7f7453c6d70a406ec14f34106d25809ce5ec7484d9 not found: ID does not exist" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.371276 4788 scope.go:117] "RemoveContainer" containerID="ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8" Oct 10 15:20:45 crc kubenswrapper[4788]: E1010 15:20:45.371674 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8\": container with ID starting with ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8 not found: ID does not exist" containerID="ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8" Oct 10 15:20:45 crc kubenswrapper[4788]: I1010 15:20:45.371720 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8"} err="failed to get container status \"ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8\": rpc error: code = NotFound desc = could not find container \"ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8\": container with ID starting with ed82cdb2b34cf959bdaa35ee2d11d4b29d470a9e3986a29140a8a4d8e7a328d8 not found: ID does not exist" Oct 10 15:20:46 crc kubenswrapper[4788]: I1010 15:20:46.251422 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" path="/var/lib/kubelet/pods/fa1f8039-d1cb-466d-9dc2-8881eeeefa36/volumes" Oct 10 15:20:59 crc kubenswrapper[4788]: I1010 15:20:59.407496 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:20:59 crc kubenswrapper[4788]: I1010 15:20:59.408408 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:20:59 crc kubenswrapper[4788]: I1010 15:20:59.408473 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:20:59 crc kubenswrapper[4788]: I1010 15:20:59.409388 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:20:59 crc kubenswrapper[4788]: I1010 15:20:59.409477 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" gracePeriod=600 Oct 10 15:20:59 crc kubenswrapper[4788]: E1010 15:20:59.548819 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:21:00 crc kubenswrapper[4788]: I1010 15:21:00.430215 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" exitCode=0 Oct 10 15:21:00 crc kubenswrapper[4788]: I1010 15:21:00.430284 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49"} Oct 10 15:21:00 crc kubenswrapper[4788]: I1010 15:21:00.430318 4788 scope.go:117] "RemoveContainer" containerID="2db1babb5b22825b16cd8f5b57897f948f0ae665e548e72a03b5b815c79d0f5d" Oct 10 15:21:00 crc kubenswrapper[4788]: I1010 15:21:00.435508 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:21:00 crc kubenswrapper[4788]: E1010 15:21:00.436201 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:21:15 crc kubenswrapper[4788]: I1010 15:21:15.234104 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:21:15 crc kubenswrapper[4788]: E1010 15:21:15.235188 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:21:29 crc kubenswrapper[4788]: I1010 15:21:29.234807 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:21:29 crc kubenswrapper[4788]: E1010 15:21:29.236131 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:21:40 crc kubenswrapper[4788]: I1010 15:21:40.234801 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:21:40 crc kubenswrapper[4788]: E1010 15:21:40.235771 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.468107 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gp8m2"] Oct 10 15:21:46 crc kubenswrapper[4788]: E1010 15:21:46.469075 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerName="extract-utilities" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.469090 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerName="extract-utilities" Oct 10 15:21:46 crc kubenswrapper[4788]: E1010 15:21:46.469098 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerName="extract-content" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.469104 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerName="extract-content" Oct 10 15:21:46 crc kubenswrapper[4788]: E1010 15:21:46.469122 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerName="registry-server" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.469129 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerName="registry-server" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.469310 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa1f8039-d1cb-466d-9dc2-8881eeeefa36" containerName="registry-server" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.470838 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.488013 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-utilities\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.488300 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chzfp\" (UniqueName: \"kubernetes.io/projected/3c45d104-443b-4afc-8c76-0bf059e1b2b8-kube-api-access-chzfp\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.488452 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-catalog-content\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.493152 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp8m2"] Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.589304 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-catalog-content\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.589403 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-utilities\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.589443 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chzfp\" (UniqueName: \"kubernetes.io/projected/3c45d104-443b-4afc-8c76-0bf059e1b2b8-kube-api-access-chzfp\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.589884 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-catalog-content\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.590014 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-utilities\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.612918 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chzfp\" (UniqueName: \"kubernetes.io/projected/3c45d104-443b-4afc-8c76-0bf059e1b2b8-kube-api-access-chzfp\") pod \"redhat-marketplace-gp8m2\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:46 crc kubenswrapper[4788]: I1010 15:21:46.799872 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:47 crc kubenswrapper[4788]: I1010 15:21:47.051765 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp8m2"] Oct 10 15:21:47 crc kubenswrapper[4788]: W1010 15:21:47.059986 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c45d104_443b_4afc_8c76_0bf059e1b2b8.slice/crio-9b39b443cf3716515472ecc1561afa144d4b6f4f1f653e022f7d17ff7463320a WatchSource:0}: Error finding container 9b39b443cf3716515472ecc1561afa144d4b6f4f1f653e022f7d17ff7463320a: Status 404 returned error can't find the container with id 9b39b443cf3716515472ecc1561afa144d4b6f4f1f653e022f7d17ff7463320a Oct 10 15:21:47 crc kubenswrapper[4788]: I1010 15:21:47.885955 4788 generic.go:334] "Generic (PLEG): container finished" podID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerID="1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b" exitCode=0 Oct 10 15:21:47 crc kubenswrapper[4788]: I1010 15:21:47.886022 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp8m2" event={"ID":"3c45d104-443b-4afc-8c76-0bf059e1b2b8","Type":"ContainerDied","Data":"1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b"} Oct 10 15:21:47 crc kubenswrapper[4788]: I1010 15:21:47.886064 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp8m2" event={"ID":"3c45d104-443b-4afc-8c76-0bf059e1b2b8","Type":"ContainerStarted","Data":"9b39b443cf3716515472ecc1561afa144d4b6f4f1f653e022f7d17ff7463320a"} Oct 10 15:21:48 crc kubenswrapper[4788]: I1010 15:21:48.900630 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp8m2" event={"ID":"3c45d104-443b-4afc-8c76-0bf059e1b2b8","Type":"ContainerStarted","Data":"9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37"} Oct 10 15:21:49 crc kubenswrapper[4788]: I1010 15:21:49.913488 4788 generic.go:334] "Generic (PLEG): container finished" podID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerID="9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37" exitCode=0 Oct 10 15:21:49 crc kubenswrapper[4788]: I1010 15:21:49.913553 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp8m2" event={"ID":"3c45d104-443b-4afc-8c76-0bf059e1b2b8","Type":"ContainerDied","Data":"9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37"} Oct 10 15:21:50 crc kubenswrapper[4788]: I1010 15:21:50.924413 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp8m2" event={"ID":"3c45d104-443b-4afc-8c76-0bf059e1b2b8","Type":"ContainerStarted","Data":"3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2"} Oct 10 15:21:50 crc kubenswrapper[4788]: I1010 15:21:50.952076 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gp8m2" podStartSLOduration=2.527696293 podStartE2EDuration="4.952054959s" podCreationTimestamp="2025-10-10 15:21:46 +0000 UTC" firstStartedPulling="2025-10-10 15:21:47.88960488 +0000 UTC m=+2210.339320448" lastFinishedPulling="2025-10-10 15:21:50.313963526 +0000 UTC m=+2212.763679114" observedRunningTime="2025-10-10 15:21:50.951809302 +0000 UTC m=+2213.401524860" watchObservedRunningTime="2025-10-10 15:21:50.952054959 +0000 UTC m=+2213.401770507" Oct 10 15:21:53 crc kubenswrapper[4788]: I1010 15:21:53.234435 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:21:53 crc kubenswrapper[4788]: E1010 15:21:53.234822 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:21:56 crc kubenswrapper[4788]: I1010 15:21:56.801082 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:56 crc kubenswrapper[4788]: I1010 15:21:56.801986 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:56 crc kubenswrapper[4788]: I1010 15:21:56.883376 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:57 crc kubenswrapper[4788]: I1010 15:21:57.050675 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:57 crc kubenswrapper[4788]: I1010 15:21:57.136548 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp8m2"] Oct 10 15:21:58 crc kubenswrapper[4788]: I1010 15:21:58.997596 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gp8m2" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerName="registry-server" containerID="cri-o://3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2" gracePeriod=2 Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.455300 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.626771 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-utilities\") pod \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.627021 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-catalog-content\") pod \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.627073 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chzfp\" (UniqueName: \"kubernetes.io/projected/3c45d104-443b-4afc-8c76-0bf059e1b2b8-kube-api-access-chzfp\") pod \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\" (UID: \"3c45d104-443b-4afc-8c76-0bf059e1b2b8\") " Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.628569 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-utilities" (OuterVolumeSpecName: "utilities") pod "3c45d104-443b-4afc-8c76-0bf059e1b2b8" (UID: "3c45d104-443b-4afc-8c76-0bf059e1b2b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.632973 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c45d104-443b-4afc-8c76-0bf059e1b2b8-kube-api-access-chzfp" (OuterVolumeSpecName: "kube-api-access-chzfp") pod "3c45d104-443b-4afc-8c76-0bf059e1b2b8" (UID: "3c45d104-443b-4afc-8c76-0bf059e1b2b8"). InnerVolumeSpecName "kube-api-access-chzfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.646196 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c45d104-443b-4afc-8c76-0bf059e1b2b8" (UID: "3c45d104-443b-4afc-8c76-0bf059e1b2b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.728783 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.728838 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c45d104-443b-4afc-8c76-0bf059e1b2b8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:21:59 crc kubenswrapper[4788]: I1010 15:21:59.728863 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chzfp\" (UniqueName: \"kubernetes.io/projected/3c45d104-443b-4afc-8c76-0bf059e1b2b8-kube-api-access-chzfp\") on node \"crc\" DevicePath \"\"" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.006869 4788 generic.go:334] "Generic (PLEG): container finished" podID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerID="3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2" exitCode=0 Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.006912 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp8m2" event={"ID":"3c45d104-443b-4afc-8c76-0bf059e1b2b8","Type":"ContainerDied","Data":"3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2"} Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.006943 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp8m2" event={"ID":"3c45d104-443b-4afc-8c76-0bf059e1b2b8","Type":"ContainerDied","Data":"9b39b443cf3716515472ecc1561afa144d4b6f4f1f653e022f7d17ff7463320a"} Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.006962 4788 scope.go:117] "RemoveContainer" containerID="3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.006976 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp8m2" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.049425 4788 scope.go:117] "RemoveContainer" containerID="9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.055117 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp8m2"] Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.061435 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp8m2"] Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.083450 4788 scope.go:117] "RemoveContainer" containerID="1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.118840 4788 scope.go:117] "RemoveContainer" containerID="3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2" Oct 10 15:22:00 crc kubenswrapper[4788]: E1010 15:22:00.119518 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2\": container with ID starting with 3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2 not found: ID does not exist" containerID="3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.119561 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2"} err="failed to get container status \"3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2\": rpc error: code = NotFound desc = could not find container \"3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2\": container with ID starting with 3854dfaa5361351cdf264d320e282f591a6ebf5bb8c57e81e9e6d993e1b307a2 not found: ID does not exist" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.119584 4788 scope.go:117] "RemoveContainer" containerID="9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37" Oct 10 15:22:00 crc kubenswrapper[4788]: E1010 15:22:00.120753 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37\": container with ID starting with 9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37 not found: ID does not exist" containerID="9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.120777 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37"} err="failed to get container status \"9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37\": rpc error: code = NotFound desc = could not find container \"9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37\": container with ID starting with 9ed517442343b1518df8dd69971ef471fbdc25a07dc1541619566e8d71b17c37 not found: ID does not exist" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.120796 4788 scope.go:117] "RemoveContainer" containerID="1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b" Oct 10 15:22:00 crc kubenswrapper[4788]: E1010 15:22:00.123366 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b\": container with ID starting with 1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b not found: ID does not exist" containerID="1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.123457 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b"} err="failed to get container status \"1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b\": rpc error: code = NotFound desc = could not find container \"1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b\": container with ID starting with 1b555418ca2b9e66164783b91b421486a02e344b4f733b4fd8e57dec957cba5b not found: ID does not exist" Oct 10 15:22:00 crc kubenswrapper[4788]: I1010 15:22:00.248455 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" path="/var/lib/kubelet/pods/3c45d104-443b-4afc-8c76-0bf059e1b2b8/volumes" Oct 10 15:22:04 crc kubenswrapper[4788]: I1010 15:22:04.234497 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:22:04 crc kubenswrapper[4788]: E1010 15:22:04.235881 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:22:18 crc kubenswrapper[4788]: I1010 15:22:18.245021 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:22:18 crc kubenswrapper[4788]: E1010 15:22:18.254474 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:22:33 crc kubenswrapper[4788]: I1010 15:22:33.233849 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:22:33 crc kubenswrapper[4788]: E1010 15:22:33.234767 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:22:44 crc kubenswrapper[4788]: I1010 15:22:44.234983 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:22:44 crc kubenswrapper[4788]: E1010 15:22:44.236081 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:22:57 crc kubenswrapper[4788]: I1010 15:22:57.233750 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:22:57 crc kubenswrapper[4788]: E1010 15:22:57.234607 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:23:09 crc kubenswrapper[4788]: I1010 15:23:09.235032 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:23:09 crc kubenswrapper[4788]: E1010 15:23:09.237582 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:23:21 crc kubenswrapper[4788]: I1010 15:23:21.234992 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:23:21 crc kubenswrapper[4788]: E1010 15:23:21.236278 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:23:32 crc kubenswrapper[4788]: I1010 15:23:32.234269 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:23:32 crc kubenswrapper[4788]: E1010 15:23:32.236062 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:23:43 crc kubenswrapper[4788]: I1010 15:23:43.233789 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:23:43 crc kubenswrapper[4788]: E1010 15:23:43.234520 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:23:54 crc kubenswrapper[4788]: I1010 15:23:54.233707 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:23:54 crc kubenswrapper[4788]: E1010 15:23:54.234527 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:24:08 crc kubenswrapper[4788]: I1010 15:24:08.238443 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:24:08 crc kubenswrapper[4788]: E1010 15:24:08.239643 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:24:19 crc kubenswrapper[4788]: I1010 15:24:19.234653 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:24:19 crc kubenswrapper[4788]: E1010 15:24:19.235747 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:24:34 crc kubenswrapper[4788]: I1010 15:24:34.233808 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:24:34 crc kubenswrapper[4788]: E1010 15:24:34.234998 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:24:48 crc kubenswrapper[4788]: I1010 15:24:48.238203 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:24:48 crc kubenswrapper[4788]: E1010 15:24:48.238994 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:25:01 crc kubenswrapper[4788]: I1010 15:25:01.233461 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:25:01 crc kubenswrapper[4788]: E1010 15:25:01.234398 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:25:16 crc kubenswrapper[4788]: I1010 15:25:16.234237 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:25:16 crc kubenswrapper[4788]: E1010 15:25:16.235123 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:25:31 crc kubenswrapper[4788]: I1010 15:25:31.234377 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:25:31 crc kubenswrapper[4788]: E1010 15:25:31.235543 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:25:45 crc kubenswrapper[4788]: I1010 15:25:45.234135 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:25:45 crc kubenswrapper[4788]: E1010 15:25:45.234850 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:25:59 crc kubenswrapper[4788]: I1010 15:25:59.234600 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:25:59 crc kubenswrapper[4788]: E1010 15:25:59.237128 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:26:13 crc kubenswrapper[4788]: I1010 15:26:13.234025 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:26:14 crc kubenswrapper[4788]: I1010 15:26:14.346458 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"173c90a1dafe6fb0d63e0f510dcf655a7f0d79894110523bf55a839ad57b281b"} Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.172829 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dlrv8"] Oct 10 15:28:07 crc kubenswrapper[4788]: E1010 15:28:07.173986 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerName="extract-content" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.174023 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerName="extract-content" Oct 10 15:28:07 crc kubenswrapper[4788]: E1010 15:28:07.174038 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerName="registry-server" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.174045 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerName="registry-server" Oct 10 15:28:07 crc kubenswrapper[4788]: E1010 15:28:07.174098 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerName="extract-utilities" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.174108 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerName="extract-utilities" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.174325 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c45d104-443b-4afc-8c76-0bf059e1b2b8" containerName="registry-server" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.179455 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.189606 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dlrv8"] Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.281120 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-catalog-content\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.281250 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-utilities\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.281329 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjkbx\" (UniqueName: \"kubernetes.io/projected/7f0043e0-5527-482b-becb-d90b131d575c-kube-api-access-gjkbx\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.383011 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-utilities\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.383121 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjkbx\" (UniqueName: \"kubernetes.io/projected/7f0043e0-5527-482b-becb-d90b131d575c-kube-api-access-gjkbx\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.383277 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-catalog-content\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.384327 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-catalog-content\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.384381 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-utilities\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.410887 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjkbx\" (UniqueName: \"kubernetes.io/projected/7f0043e0-5527-482b-becb-d90b131d575c-kube-api-access-gjkbx\") pod \"community-operators-dlrv8\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.504182 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:07 crc kubenswrapper[4788]: I1010 15:28:07.984313 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dlrv8"] Oct 10 15:28:08 crc kubenswrapper[4788]: I1010 15:28:08.406381 4788 generic.go:334] "Generic (PLEG): container finished" podID="7f0043e0-5527-482b-becb-d90b131d575c" containerID="130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa" exitCode=0 Oct 10 15:28:08 crc kubenswrapper[4788]: I1010 15:28:08.406435 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlrv8" event={"ID":"7f0043e0-5527-482b-becb-d90b131d575c","Type":"ContainerDied","Data":"130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa"} Oct 10 15:28:08 crc kubenswrapper[4788]: I1010 15:28:08.406794 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlrv8" event={"ID":"7f0043e0-5527-482b-becb-d90b131d575c","Type":"ContainerStarted","Data":"c7639925f6da1c411355a05951bdc57e88c9459fa3cc4c0e13cbb8f28c91083b"} Oct 10 15:28:08 crc kubenswrapper[4788]: I1010 15:28:08.408711 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 15:28:09 crc kubenswrapper[4788]: I1010 15:28:09.418720 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlrv8" event={"ID":"7f0043e0-5527-482b-becb-d90b131d575c","Type":"ContainerStarted","Data":"2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99"} Oct 10 15:28:09 crc kubenswrapper[4788]: E1010 15:28:09.630525 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f0043e0_5527_482b_becb_d90b131d575c.slice/crio-conmon-2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f0043e0_5527_482b_becb_d90b131d575c.slice/crio-2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99.scope\": RecentStats: unable to find data in memory cache]" Oct 10 15:28:10 crc kubenswrapper[4788]: I1010 15:28:10.428660 4788 generic.go:334] "Generic (PLEG): container finished" podID="7f0043e0-5527-482b-becb-d90b131d575c" containerID="2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99" exitCode=0 Oct 10 15:28:10 crc kubenswrapper[4788]: I1010 15:28:10.428808 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlrv8" event={"ID":"7f0043e0-5527-482b-becb-d90b131d575c","Type":"ContainerDied","Data":"2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99"} Oct 10 15:28:11 crc kubenswrapper[4788]: I1010 15:28:11.443526 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlrv8" event={"ID":"7f0043e0-5527-482b-becb-d90b131d575c","Type":"ContainerStarted","Data":"8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9"} Oct 10 15:28:11 crc kubenswrapper[4788]: I1010 15:28:11.468468 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dlrv8" podStartSLOduration=1.672261036 podStartE2EDuration="4.468439929s" podCreationTimestamp="2025-10-10 15:28:07 +0000 UTC" firstStartedPulling="2025-10-10 15:28:08.408438137 +0000 UTC m=+2590.858153685" lastFinishedPulling="2025-10-10 15:28:11.20461699 +0000 UTC m=+2593.654332578" observedRunningTime="2025-10-10 15:28:11.46550192 +0000 UTC m=+2593.915217488" watchObservedRunningTime="2025-10-10 15:28:11.468439929 +0000 UTC m=+2593.918155477" Oct 10 15:28:17 crc kubenswrapper[4788]: I1010 15:28:17.504941 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:17 crc kubenswrapper[4788]: I1010 15:28:17.505989 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:17 crc kubenswrapper[4788]: I1010 15:28:17.567165 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:18 crc kubenswrapper[4788]: I1010 15:28:18.578532 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:18 crc kubenswrapper[4788]: I1010 15:28:18.650480 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dlrv8"] Oct 10 15:28:20 crc kubenswrapper[4788]: I1010 15:28:20.524840 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dlrv8" podUID="7f0043e0-5527-482b-becb-d90b131d575c" containerName="registry-server" containerID="cri-o://8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9" gracePeriod=2 Oct 10 15:28:20 crc kubenswrapper[4788]: I1010 15:28:20.945547 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.105420 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjkbx\" (UniqueName: \"kubernetes.io/projected/7f0043e0-5527-482b-becb-d90b131d575c-kube-api-access-gjkbx\") pod \"7f0043e0-5527-482b-becb-d90b131d575c\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.105499 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-utilities\") pod \"7f0043e0-5527-482b-becb-d90b131d575c\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.105597 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-catalog-content\") pod \"7f0043e0-5527-482b-becb-d90b131d575c\" (UID: \"7f0043e0-5527-482b-becb-d90b131d575c\") " Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.107490 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-utilities" (OuterVolumeSpecName: "utilities") pod "7f0043e0-5527-482b-becb-d90b131d575c" (UID: "7f0043e0-5527-482b-becb-d90b131d575c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.116838 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f0043e0-5527-482b-becb-d90b131d575c-kube-api-access-gjkbx" (OuterVolumeSpecName: "kube-api-access-gjkbx") pod "7f0043e0-5527-482b-becb-d90b131d575c" (UID: "7f0043e0-5527-482b-becb-d90b131d575c"). InnerVolumeSpecName "kube-api-access-gjkbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.162923 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f0043e0-5527-482b-becb-d90b131d575c" (UID: "7f0043e0-5527-482b-becb-d90b131d575c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.206764 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjkbx\" (UniqueName: \"kubernetes.io/projected/7f0043e0-5527-482b-becb-d90b131d575c-kube-api-access-gjkbx\") on node \"crc\" DevicePath \"\"" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.206803 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.206813 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f0043e0-5527-482b-becb-d90b131d575c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.541877 4788 generic.go:334] "Generic (PLEG): container finished" podID="7f0043e0-5527-482b-becb-d90b131d575c" containerID="8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9" exitCode=0 Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.541936 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlrv8" event={"ID":"7f0043e0-5527-482b-becb-d90b131d575c","Type":"ContainerDied","Data":"8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9"} Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.541985 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dlrv8" event={"ID":"7f0043e0-5527-482b-becb-d90b131d575c","Type":"ContainerDied","Data":"c7639925f6da1c411355a05951bdc57e88c9459fa3cc4c0e13cbb8f28c91083b"} Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.541990 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dlrv8" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.542054 4788 scope.go:117] "RemoveContainer" containerID="8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.584291 4788 scope.go:117] "RemoveContainer" containerID="2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.590628 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dlrv8"] Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.601299 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dlrv8"] Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.636491 4788 scope.go:117] "RemoveContainer" containerID="130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.660637 4788 scope.go:117] "RemoveContainer" containerID="8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9" Oct 10 15:28:21 crc kubenswrapper[4788]: E1010 15:28:21.661191 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9\": container with ID starting with 8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9 not found: ID does not exist" containerID="8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.661225 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9"} err="failed to get container status \"8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9\": rpc error: code = NotFound desc = could not find container \"8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9\": container with ID starting with 8b7fe8b92e96710a07cab72a6d97a92cd146974c5569c19faa270612621e23f9 not found: ID does not exist" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.661248 4788 scope.go:117] "RemoveContainer" containerID="2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99" Oct 10 15:28:21 crc kubenswrapper[4788]: E1010 15:28:21.661776 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99\": container with ID starting with 2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99 not found: ID does not exist" containerID="2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.661879 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99"} err="failed to get container status \"2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99\": rpc error: code = NotFound desc = could not find container \"2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99\": container with ID starting with 2c10f04fe4a96bab3fa98185915328b76b3f749671bcb3ea7d634304cbb67e99 not found: ID does not exist" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.661940 4788 scope.go:117] "RemoveContainer" containerID="130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa" Oct 10 15:28:21 crc kubenswrapper[4788]: E1010 15:28:21.662457 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa\": container with ID starting with 130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa not found: ID does not exist" containerID="130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa" Oct 10 15:28:21 crc kubenswrapper[4788]: I1010 15:28:21.662534 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa"} err="failed to get container status \"130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa\": rpc error: code = NotFound desc = could not find container \"130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa\": container with ID starting with 130b3ef581016c7eef6409e82ba44c10810fbe4f3d59dcbb2be50180eb09e3fa not found: ID does not exist" Oct 10 15:28:22 crc kubenswrapper[4788]: I1010 15:28:22.254170 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f0043e0-5527-482b-becb-d90b131d575c" path="/var/lib/kubelet/pods/7f0043e0-5527-482b-becb-d90b131d575c/volumes" Oct 10 15:28:29 crc kubenswrapper[4788]: I1010 15:28:29.406936 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:28:29 crc kubenswrapper[4788]: I1010 15:28:29.407692 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:28:59 crc kubenswrapper[4788]: I1010 15:28:59.406585 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:28:59 crc kubenswrapper[4788]: I1010 15:28:59.408061 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:29:29 crc kubenswrapper[4788]: I1010 15:29:29.405808 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:29:29 crc kubenswrapper[4788]: I1010 15:29:29.406800 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:29:29 crc kubenswrapper[4788]: I1010 15:29:29.406868 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:29:29 crc kubenswrapper[4788]: I1010 15:29:29.407867 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"173c90a1dafe6fb0d63e0f510dcf655a7f0d79894110523bf55a839ad57b281b"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:29:29 crc kubenswrapper[4788]: I1010 15:29:29.407945 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://173c90a1dafe6fb0d63e0f510dcf655a7f0d79894110523bf55a839ad57b281b" gracePeriod=600 Oct 10 15:29:30 crc kubenswrapper[4788]: I1010 15:29:30.166910 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="173c90a1dafe6fb0d63e0f510dcf655a7f0d79894110523bf55a839ad57b281b" exitCode=0 Oct 10 15:29:30 crc kubenswrapper[4788]: I1010 15:29:30.167546 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"173c90a1dafe6fb0d63e0f510dcf655a7f0d79894110523bf55a839ad57b281b"} Oct 10 15:29:30 crc kubenswrapper[4788]: I1010 15:29:30.167607 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51"} Oct 10 15:29:30 crc kubenswrapper[4788]: I1010 15:29:30.167641 4788 scope.go:117] "RemoveContainer" containerID="ab5581cb784f29018bcd166a285b4560af5559a0f081b87a187fd7a259d40b49" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.169544 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn"] Oct 10 15:30:00 crc kubenswrapper[4788]: E1010 15:30:00.170579 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0043e0-5527-482b-becb-d90b131d575c" containerName="extract-utilities" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.170597 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0043e0-5527-482b-becb-d90b131d575c" containerName="extract-utilities" Oct 10 15:30:00 crc kubenswrapper[4788]: E1010 15:30:00.170650 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0043e0-5527-482b-becb-d90b131d575c" containerName="extract-content" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.170659 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0043e0-5527-482b-becb-d90b131d575c" containerName="extract-content" Oct 10 15:30:00 crc kubenswrapper[4788]: E1010 15:30:00.170669 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0043e0-5527-482b-becb-d90b131d575c" containerName="registry-server" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.170679 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0043e0-5527-482b-becb-d90b131d575c" containerName="registry-server" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.170880 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f0043e0-5527-482b-becb-d90b131d575c" containerName="registry-server" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.171510 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.176715 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.176827 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.186794 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn"] Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.304768 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aab10930-60df-4195-b105-da165a8b4fea-config-volume\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.304821 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aab10930-60df-4195-b105-da165a8b4fea-secret-volume\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.304842 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5q9b\" (UniqueName: \"kubernetes.io/projected/aab10930-60df-4195-b105-da165a8b4fea-kube-api-access-z5q9b\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.407275 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aab10930-60df-4195-b105-da165a8b4fea-config-volume\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.407339 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aab10930-60df-4195-b105-da165a8b4fea-secret-volume\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.407369 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5q9b\" (UniqueName: \"kubernetes.io/projected/aab10930-60df-4195-b105-da165a8b4fea-kube-api-access-z5q9b\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.408428 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aab10930-60df-4195-b105-da165a8b4fea-config-volume\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.417628 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aab10930-60df-4195-b105-da165a8b4fea-secret-volume\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.429400 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5q9b\" (UniqueName: \"kubernetes.io/projected/aab10930-60df-4195-b105-da165a8b4fea-kube-api-access-z5q9b\") pod \"collect-profiles-29335170-px4wn\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.498993 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:00 crc kubenswrapper[4788]: I1010 15:30:00.980003 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn"] Oct 10 15:30:01 crc kubenswrapper[4788]: I1010 15:30:01.449935 4788 generic.go:334] "Generic (PLEG): container finished" podID="aab10930-60df-4195-b105-da165a8b4fea" containerID="9b000249408e6ca31216e6a3d236e835b3baf6337604721730d3c7be9c80637f" exitCode=0 Oct 10 15:30:01 crc kubenswrapper[4788]: I1010 15:30:01.450166 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" event={"ID":"aab10930-60df-4195-b105-da165a8b4fea","Type":"ContainerDied","Data":"9b000249408e6ca31216e6a3d236e835b3baf6337604721730d3c7be9c80637f"} Oct 10 15:30:01 crc kubenswrapper[4788]: I1010 15:30:01.450274 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" event={"ID":"aab10930-60df-4195-b105-da165a8b4fea","Type":"ContainerStarted","Data":"4be207b329fb59bdb46a10cef318c5cd10c42611883f87d7978a5b95d8a8870b"} Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.755118 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.879620 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5q9b\" (UniqueName: \"kubernetes.io/projected/aab10930-60df-4195-b105-da165a8b4fea-kube-api-access-z5q9b\") pod \"aab10930-60df-4195-b105-da165a8b4fea\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.881950 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aab10930-60df-4195-b105-da165a8b4fea-config-volume\") pod \"aab10930-60df-4195-b105-da165a8b4fea\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.882051 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aab10930-60df-4195-b105-da165a8b4fea-secret-volume\") pod \"aab10930-60df-4195-b105-da165a8b4fea\" (UID: \"aab10930-60df-4195-b105-da165a8b4fea\") " Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.882810 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aab10930-60df-4195-b105-da165a8b4fea-config-volume" (OuterVolumeSpecName: "config-volume") pod "aab10930-60df-4195-b105-da165a8b4fea" (UID: "aab10930-60df-4195-b105-da165a8b4fea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.885979 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab10930-60df-4195-b105-da165a8b4fea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aab10930-60df-4195-b105-da165a8b4fea" (UID: "aab10930-60df-4195-b105-da165a8b4fea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.886083 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aab10930-60df-4195-b105-da165a8b4fea-kube-api-access-z5q9b" (OuterVolumeSpecName: "kube-api-access-z5q9b") pod "aab10930-60df-4195-b105-da165a8b4fea" (UID: "aab10930-60df-4195-b105-da165a8b4fea"). InnerVolumeSpecName "kube-api-access-z5q9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.984013 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aab10930-60df-4195-b105-da165a8b4fea-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.984425 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aab10930-60df-4195-b105-da165a8b4fea-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 15:30:02 crc kubenswrapper[4788]: I1010 15:30:02.984438 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5q9b\" (UniqueName: \"kubernetes.io/projected/aab10930-60df-4195-b105-da165a8b4fea-kube-api-access-z5q9b\") on node \"crc\" DevicePath \"\"" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.436005 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-58n98"] Oct 10 15:30:03 crc kubenswrapper[4788]: E1010 15:30:03.436320 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab10930-60df-4195-b105-da165a8b4fea" containerName="collect-profiles" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.436332 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab10930-60df-4195-b105-da165a8b4fea" containerName="collect-profiles" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.436472 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab10930-60df-4195-b105-da165a8b4fea" containerName="collect-profiles" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.437402 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.451382 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-58n98"] Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.470675 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" event={"ID":"aab10930-60df-4195-b105-da165a8b4fea","Type":"ContainerDied","Data":"4be207b329fb59bdb46a10cef318c5cd10c42611883f87d7978a5b95d8a8870b"} Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.470721 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4be207b329fb59bdb46a10cef318c5cd10c42611883f87d7978a5b95d8a8870b" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.470788 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.591499 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-catalog-content\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.591560 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw78s\" (UniqueName: \"kubernetes.io/projected/7b9db77f-3dc4-4367-9a52-d5815969c44d-kube-api-access-nw78s\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.591589 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-utilities\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.692923 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-catalog-content\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.692978 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw78s\" (UniqueName: \"kubernetes.io/projected/7b9db77f-3dc4-4367-9a52-d5815969c44d-kube-api-access-nw78s\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.693006 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-utilities\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.693456 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-catalog-content\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.693519 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-utilities\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.717421 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw78s\" (UniqueName: \"kubernetes.io/projected/7b9db77f-3dc4-4367-9a52-d5815969c44d-kube-api-access-nw78s\") pod \"redhat-operators-58n98\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.775762 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.830849 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r"] Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.841315 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335125-24p8r"] Oct 10 15:30:03 crc kubenswrapper[4788]: I1010 15:30:03.984082 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-58n98"] Oct 10 15:30:03 crc kubenswrapper[4788]: W1010 15:30:03.995809 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b9db77f_3dc4_4367_9a52_d5815969c44d.slice/crio-8f2db8944b02af20483e46d34fb125d2573b07ba9212c89520ae3038f6dfa5fb WatchSource:0}: Error finding container 8f2db8944b02af20483e46d34fb125d2573b07ba9212c89520ae3038f6dfa5fb: Status 404 returned error can't find the container with id 8f2db8944b02af20483e46d34fb125d2573b07ba9212c89520ae3038f6dfa5fb Oct 10 15:30:04 crc kubenswrapper[4788]: I1010 15:30:04.244294 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bf9df01-bc0d-46cf-8786-0ab3acea7a7a" path="/var/lib/kubelet/pods/2bf9df01-bc0d-46cf-8786-0ab3acea7a7a/volumes" Oct 10 15:30:04 crc kubenswrapper[4788]: I1010 15:30:04.479229 4788 generic.go:334] "Generic (PLEG): container finished" podID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerID="ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180" exitCode=0 Oct 10 15:30:04 crc kubenswrapper[4788]: I1010 15:30:04.479277 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58n98" event={"ID":"7b9db77f-3dc4-4367-9a52-d5815969c44d","Type":"ContainerDied","Data":"ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180"} Oct 10 15:30:04 crc kubenswrapper[4788]: I1010 15:30:04.479306 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58n98" event={"ID":"7b9db77f-3dc4-4367-9a52-d5815969c44d","Type":"ContainerStarted","Data":"8f2db8944b02af20483e46d34fb125d2573b07ba9212c89520ae3038f6dfa5fb"} Oct 10 15:30:05 crc kubenswrapper[4788]: I1010 15:30:05.490661 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58n98" event={"ID":"7b9db77f-3dc4-4367-9a52-d5815969c44d","Type":"ContainerStarted","Data":"784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7"} Oct 10 15:30:06 crc kubenswrapper[4788]: I1010 15:30:06.503737 4788 generic.go:334] "Generic (PLEG): container finished" podID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerID="784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7" exitCode=0 Oct 10 15:30:06 crc kubenswrapper[4788]: I1010 15:30:06.503806 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58n98" event={"ID":"7b9db77f-3dc4-4367-9a52-d5815969c44d","Type":"ContainerDied","Data":"784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7"} Oct 10 15:30:07 crc kubenswrapper[4788]: I1010 15:30:07.517271 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58n98" event={"ID":"7b9db77f-3dc4-4367-9a52-d5815969c44d","Type":"ContainerStarted","Data":"649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927"} Oct 10 15:30:07 crc kubenswrapper[4788]: I1010 15:30:07.554277 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-58n98" podStartSLOduration=1.912914973 podStartE2EDuration="4.554248008s" podCreationTimestamp="2025-10-10 15:30:03 +0000 UTC" firstStartedPulling="2025-10-10 15:30:04.481100343 +0000 UTC m=+2706.930815891" lastFinishedPulling="2025-10-10 15:30:07.122433368 +0000 UTC m=+2709.572148926" observedRunningTime="2025-10-10 15:30:07.549741357 +0000 UTC m=+2709.999456945" watchObservedRunningTime="2025-10-10 15:30:07.554248008 +0000 UTC m=+2710.003963576" Oct 10 15:30:10 crc kubenswrapper[4788]: I1010 15:30:10.761000 4788 scope.go:117] "RemoveContainer" containerID="95d45dca47c61283fef4a79ffa78df851c6de47f8f641dff6f172630475ac1cd" Oct 10 15:30:13 crc kubenswrapper[4788]: I1010 15:30:13.776671 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:13 crc kubenswrapper[4788]: I1010 15:30:13.777251 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:13 crc kubenswrapper[4788]: I1010 15:30:13.820065 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:14 crc kubenswrapper[4788]: I1010 15:30:14.624582 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:14 crc kubenswrapper[4788]: I1010 15:30:14.671325 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-58n98"] Oct 10 15:30:16 crc kubenswrapper[4788]: I1010 15:30:16.594504 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-58n98" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerName="registry-server" containerID="cri-o://649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927" gracePeriod=2 Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.563275 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.610644 4788 generic.go:334] "Generic (PLEG): container finished" podID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerID="649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927" exitCode=0 Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.610708 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58n98" event={"ID":"7b9db77f-3dc4-4367-9a52-d5815969c44d","Type":"ContainerDied","Data":"649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927"} Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.610746 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-58n98" event={"ID":"7b9db77f-3dc4-4367-9a52-d5815969c44d","Type":"ContainerDied","Data":"8f2db8944b02af20483e46d34fb125d2573b07ba9212c89520ae3038f6dfa5fb"} Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.610768 4788 scope.go:117] "RemoveContainer" containerID="649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.610940 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-58n98" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.627520 4788 scope.go:117] "RemoveContainer" containerID="784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.642779 4788 scope.go:117] "RemoveContainer" containerID="ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.666901 4788 scope.go:117] "RemoveContainer" containerID="649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927" Oct 10 15:30:17 crc kubenswrapper[4788]: E1010 15:30:17.667385 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927\": container with ID starting with 649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927 not found: ID does not exist" containerID="649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.667424 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927"} err="failed to get container status \"649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927\": rpc error: code = NotFound desc = could not find container \"649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927\": container with ID starting with 649ae9cf21fd99078d94c498e1bc2534cf9fb24f0c4044b6198e07a9d6d97927 not found: ID does not exist" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.667457 4788 scope.go:117] "RemoveContainer" containerID="784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7" Oct 10 15:30:17 crc kubenswrapper[4788]: E1010 15:30:17.667720 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7\": container with ID starting with 784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7 not found: ID does not exist" containerID="784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.667844 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7"} err="failed to get container status \"784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7\": rpc error: code = NotFound desc = could not find container \"784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7\": container with ID starting with 784b1f9a7720163bd6e20fe4f24b4979c158685016fa80d25f35ae055b1336f7 not found: ID does not exist" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.667882 4788 scope.go:117] "RemoveContainer" containerID="ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180" Oct 10 15:30:17 crc kubenswrapper[4788]: E1010 15:30:17.668209 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180\": container with ID starting with ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180 not found: ID does not exist" containerID="ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.668247 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180"} err="failed to get container status \"ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180\": rpc error: code = NotFound desc = could not find container \"ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180\": container with ID starting with ab1fff2c94ed56b350eb92da4a422855cf36cd635a1c6cf37705bcd297d94180 not found: ID does not exist" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.709399 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-utilities\") pod \"7b9db77f-3dc4-4367-9a52-d5815969c44d\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.709481 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw78s\" (UniqueName: \"kubernetes.io/projected/7b9db77f-3dc4-4367-9a52-d5815969c44d-kube-api-access-nw78s\") pod \"7b9db77f-3dc4-4367-9a52-d5815969c44d\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.709518 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-catalog-content\") pod \"7b9db77f-3dc4-4367-9a52-d5815969c44d\" (UID: \"7b9db77f-3dc4-4367-9a52-d5815969c44d\") " Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.710508 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-utilities" (OuterVolumeSpecName: "utilities") pod "7b9db77f-3dc4-4367-9a52-d5815969c44d" (UID: "7b9db77f-3dc4-4367-9a52-d5815969c44d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.715368 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b9db77f-3dc4-4367-9a52-d5815969c44d-kube-api-access-nw78s" (OuterVolumeSpecName: "kube-api-access-nw78s") pod "7b9db77f-3dc4-4367-9a52-d5815969c44d" (UID: "7b9db77f-3dc4-4367-9a52-d5815969c44d"). InnerVolumeSpecName "kube-api-access-nw78s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.795045 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b9db77f-3dc4-4367-9a52-d5815969c44d" (UID: "7b9db77f-3dc4-4367-9a52-d5815969c44d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.810694 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.810728 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw78s\" (UniqueName: \"kubernetes.io/projected/7b9db77f-3dc4-4367-9a52-d5815969c44d-kube-api-access-nw78s\") on node \"crc\" DevicePath \"\"" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.810741 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9db77f-3dc4-4367-9a52-d5815969c44d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.954018 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-58n98"] Oct 10 15:30:17 crc kubenswrapper[4788]: I1010 15:30:17.958489 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-58n98"] Oct 10 15:30:18 crc kubenswrapper[4788]: I1010 15:30:18.242628 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" path="/var/lib/kubelet/pods/7b9db77f-3dc4-4367-9a52-d5815969c44d/volumes" Oct 10 15:30:58 crc kubenswrapper[4788]: I1010 15:30:58.826427 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kbszr"] Oct 10 15:30:58 crc kubenswrapper[4788]: E1010 15:30:58.827840 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerName="extract-utilities" Oct 10 15:30:58 crc kubenswrapper[4788]: I1010 15:30:58.827877 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerName="extract-utilities" Oct 10 15:30:58 crc kubenswrapper[4788]: E1010 15:30:58.827908 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerName="registry-server" Oct 10 15:30:58 crc kubenswrapper[4788]: I1010 15:30:58.827927 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerName="registry-server" Oct 10 15:30:58 crc kubenswrapper[4788]: E1010 15:30:58.827978 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerName="extract-content" Oct 10 15:30:58 crc kubenswrapper[4788]: I1010 15:30:58.827997 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerName="extract-content" Oct 10 15:30:58 crc kubenswrapper[4788]: I1010 15:30:58.828397 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b9db77f-3dc4-4367-9a52-d5815969c44d" containerName="registry-server" Oct 10 15:30:58 crc kubenswrapper[4788]: I1010 15:30:58.830827 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:58 crc kubenswrapper[4788]: I1010 15:30:58.843204 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbszr"] Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.003039 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-catalog-content\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.003167 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-utilities\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.003250 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdcdt\" (UniqueName: \"kubernetes.io/projected/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-kube-api-access-jdcdt\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.105242 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-catalog-content\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.105322 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-utilities\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.105367 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdcdt\" (UniqueName: \"kubernetes.io/projected/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-kube-api-access-jdcdt\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.105773 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-catalog-content\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.106019 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-utilities\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.132385 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdcdt\" (UniqueName: \"kubernetes.io/projected/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-kube-api-access-jdcdt\") pod \"certified-operators-kbszr\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.152859 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:30:59 crc kubenswrapper[4788]: I1010 15:30:59.406149 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbszr"] Oct 10 15:31:00 crc kubenswrapper[4788]: I1010 15:31:00.024695 4788 generic.go:334] "Generic (PLEG): container finished" podID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerID="443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa" exitCode=0 Oct 10 15:31:00 crc kubenswrapper[4788]: I1010 15:31:00.024763 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbszr" event={"ID":"bf695c0b-f31e-47a5-a02f-22a839f3f0f6","Type":"ContainerDied","Data":"443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa"} Oct 10 15:31:00 crc kubenswrapper[4788]: I1010 15:31:00.024821 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbszr" event={"ID":"bf695c0b-f31e-47a5-a02f-22a839f3f0f6","Type":"ContainerStarted","Data":"9c4d4b8907a2db645955630aa64400158adbdf61bb19a2f51d72c85e15ac2c7b"} Oct 10 15:31:01 crc kubenswrapper[4788]: I1010 15:31:01.035172 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbszr" event={"ID":"bf695c0b-f31e-47a5-a02f-22a839f3f0f6","Type":"ContainerStarted","Data":"9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb"} Oct 10 15:31:02 crc kubenswrapper[4788]: I1010 15:31:02.044339 4788 generic.go:334] "Generic (PLEG): container finished" podID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerID="9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb" exitCode=0 Oct 10 15:31:02 crc kubenswrapper[4788]: I1010 15:31:02.044401 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbszr" event={"ID":"bf695c0b-f31e-47a5-a02f-22a839f3f0f6","Type":"ContainerDied","Data":"9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb"} Oct 10 15:31:03 crc kubenswrapper[4788]: I1010 15:31:03.055877 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbszr" event={"ID":"bf695c0b-f31e-47a5-a02f-22a839f3f0f6","Type":"ContainerStarted","Data":"96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b"} Oct 10 15:31:03 crc kubenswrapper[4788]: I1010 15:31:03.086578 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kbszr" podStartSLOduration=2.503922074 podStartE2EDuration="5.086561565s" podCreationTimestamp="2025-10-10 15:30:58 +0000 UTC" firstStartedPulling="2025-10-10 15:31:00.026748008 +0000 UTC m=+2762.476463556" lastFinishedPulling="2025-10-10 15:31:02.609387459 +0000 UTC m=+2765.059103047" observedRunningTime="2025-10-10 15:31:03.080863373 +0000 UTC m=+2765.530578921" watchObservedRunningTime="2025-10-10 15:31:03.086561565 +0000 UTC m=+2765.536277113" Oct 10 15:31:09 crc kubenswrapper[4788]: I1010 15:31:09.153540 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:31:09 crc kubenswrapper[4788]: I1010 15:31:09.154323 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:31:09 crc kubenswrapper[4788]: I1010 15:31:09.212335 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:31:10 crc kubenswrapper[4788]: I1010 15:31:10.205931 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:31:10 crc kubenswrapper[4788]: I1010 15:31:10.281355 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kbszr"] Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.147509 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kbszr" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerName="registry-server" containerID="cri-o://96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b" gracePeriod=2 Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.579026 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.737852 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdcdt\" (UniqueName: \"kubernetes.io/projected/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-kube-api-access-jdcdt\") pod \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.738110 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-utilities\") pod \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.738359 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-catalog-content\") pod \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\" (UID: \"bf695c0b-f31e-47a5-a02f-22a839f3f0f6\") " Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.740064 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-utilities" (OuterVolumeSpecName: "utilities") pod "bf695c0b-f31e-47a5-a02f-22a839f3f0f6" (UID: "bf695c0b-f31e-47a5-a02f-22a839f3f0f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.753423 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-kube-api-access-jdcdt" (OuterVolumeSpecName: "kube-api-access-jdcdt") pod "bf695c0b-f31e-47a5-a02f-22a839f3f0f6" (UID: "bf695c0b-f31e-47a5-a02f-22a839f3f0f6"). InnerVolumeSpecName "kube-api-access-jdcdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.803017 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf695c0b-f31e-47a5-a02f-22a839f3f0f6" (UID: "bf695c0b-f31e-47a5-a02f-22a839f3f0f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.841079 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdcdt\" (UniqueName: \"kubernetes.io/projected/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-kube-api-access-jdcdt\") on node \"crc\" DevicePath \"\"" Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.841163 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:31:12 crc kubenswrapper[4788]: I1010 15:31:12.841187 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf695c0b-f31e-47a5-a02f-22a839f3f0f6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.160649 4788 generic.go:334] "Generic (PLEG): container finished" podID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerID="96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b" exitCode=0 Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.160715 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbszr" event={"ID":"bf695c0b-f31e-47a5-a02f-22a839f3f0f6","Type":"ContainerDied","Data":"96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b"} Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.162793 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbszr" event={"ID":"bf695c0b-f31e-47a5-a02f-22a839f3f0f6","Type":"ContainerDied","Data":"9c4d4b8907a2db645955630aa64400158adbdf61bb19a2f51d72c85e15ac2c7b"} Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.160764 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbszr" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.162867 4788 scope.go:117] "RemoveContainer" containerID="96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.197842 4788 scope.go:117] "RemoveContainer" containerID="9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.218542 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kbszr"] Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.233642 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kbszr"] Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.250512 4788 scope.go:117] "RemoveContainer" containerID="443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.270130 4788 scope.go:117] "RemoveContainer" containerID="96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b" Oct 10 15:31:13 crc kubenswrapper[4788]: E1010 15:31:13.270886 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b\": container with ID starting with 96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b not found: ID does not exist" containerID="96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.270959 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b"} err="failed to get container status \"96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b\": rpc error: code = NotFound desc = could not find container \"96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b\": container with ID starting with 96d9f3ea3d619167708c4ae5acfd9d30c263de1d08f4fb482981f0b193ec086b not found: ID does not exist" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.271002 4788 scope.go:117] "RemoveContainer" containerID="9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb" Oct 10 15:31:13 crc kubenswrapper[4788]: E1010 15:31:13.272100 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb\": container with ID starting with 9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb not found: ID does not exist" containerID="9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.272176 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb"} err="failed to get container status \"9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb\": rpc error: code = NotFound desc = could not find container \"9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb\": container with ID starting with 9357533be91f91b91cceb72af9a7c8572e6c03f053157692e52f44d59b1435fb not found: ID does not exist" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.272211 4788 scope.go:117] "RemoveContainer" containerID="443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa" Oct 10 15:31:13 crc kubenswrapper[4788]: E1010 15:31:13.273563 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa\": container with ID starting with 443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa not found: ID does not exist" containerID="443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa" Oct 10 15:31:13 crc kubenswrapper[4788]: I1010 15:31:13.273641 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa"} err="failed to get container status \"443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa\": rpc error: code = NotFound desc = could not find container \"443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa\": container with ID starting with 443295de79e5d87f2c776fad87b998f76676945e45b8542177f177a1d0da77fa not found: ID does not exist" Oct 10 15:31:14 crc kubenswrapper[4788]: I1010 15:31:14.249347 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" path="/var/lib/kubelet/pods/bf695c0b-f31e-47a5-a02f-22a839f3f0f6/volumes" Oct 10 15:31:29 crc kubenswrapper[4788]: I1010 15:31:29.406181 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:31:29 crc kubenswrapper[4788]: I1010 15:31:29.406917 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:31:59 crc kubenswrapper[4788]: I1010 15:31:59.406280 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:31:59 crc kubenswrapper[4788]: I1010 15:31:59.408929 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.406348 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.407372 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.407452 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.408396 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.408504 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" gracePeriod=600 Oct 10 15:32:29 crc kubenswrapper[4788]: E1010 15:32:29.554267 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.904309 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" exitCode=0 Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.904373 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51"} Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.904450 4788 scope.go:117] "RemoveContainer" containerID="173c90a1dafe6fb0d63e0f510dcf655a7f0d79894110523bf55a839ad57b281b" Oct 10 15:32:29 crc kubenswrapper[4788]: I1010 15:32:29.905252 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:32:29 crc kubenswrapper[4788]: E1010 15:32:29.905741 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:32:43 crc kubenswrapper[4788]: I1010 15:32:43.234693 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:32:43 crc kubenswrapper[4788]: E1010 15:32:43.235759 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.189121 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cm877"] Oct 10 15:32:53 crc kubenswrapper[4788]: E1010 15:32:53.189934 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerName="extract-content" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.189950 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerName="extract-content" Oct 10 15:32:53 crc kubenswrapper[4788]: E1010 15:32:53.189960 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerName="registry-server" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.189966 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerName="registry-server" Oct 10 15:32:53 crc kubenswrapper[4788]: E1010 15:32:53.189977 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerName="extract-utilities" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.189983 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerName="extract-utilities" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.190127 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf695c0b-f31e-47a5-a02f-22a839f3f0f6" containerName="registry-server" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.191276 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.202816 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm877"] Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.355219 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfhjt\" (UniqueName: \"kubernetes.io/projected/df9de94a-cbdd-48fc-9961-a13710685573-kube-api-access-nfhjt\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.355391 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-catalog-content\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.355593 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-utilities\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.456967 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-utilities\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.457019 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfhjt\" (UniqueName: \"kubernetes.io/projected/df9de94a-cbdd-48fc-9961-a13710685573-kube-api-access-nfhjt\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.457072 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-catalog-content\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.457601 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-utilities\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.457643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-catalog-content\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.478099 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfhjt\" (UniqueName: \"kubernetes.io/projected/df9de94a-cbdd-48fc-9961-a13710685573-kube-api-access-nfhjt\") pod \"redhat-marketplace-cm877\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.570885 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:32:53 crc kubenswrapper[4788]: I1010 15:32:53.993304 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm877"] Oct 10 15:32:54 crc kubenswrapper[4788]: I1010 15:32:54.168583 4788 generic.go:334] "Generic (PLEG): container finished" podID="df9de94a-cbdd-48fc-9961-a13710685573" containerID="4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81" exitCode=0 Oct 10 15:32:54 crc kubenswrapper[4788]: I1010 15:32:54.168642 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm877" event={"ID":"df9de94a-cbdd-48fc-9961-a13710685573","Type":"ContainerDied","Data":"4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81"} Oct 10 15:32:54 crc kubenswrapper[4788]: I1010 15:32:54.168692 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm877" event={"ID":"df9de94a-cbdd-48fc-9961-a13710685573","Type":"ContainerStarted","Data":"a571642efc39fcc335e5977d503c628c42c5519d81652ca125385ea87d86e586"} Oct 10 15:32:55 crc kubenswrapper[4788]: I1010 15:32:55.178110 4788 generic.go:334] "Generic (PLEG): container finished" podID="df9de94a-cbdd-48fc-9961-a13710685573" containerID="de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e" exitCode=0 Oct 10 15:32:55 crc kubenswrapper[4788]: I1010 15:32:55.178195 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm877" event={"ID":"df9de94a-cbdd-48fc-9961-a13710685573","Type":"ContainerDied","Data":"de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e"} Oct 10 15:32:56 crc kubenswrapper[4788]: I1010 15:32:56.190648 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm877" event={"ID":"df9de94a-cbdd-48fc-9961-a13710685573","Type":"ContainerStarted","Data":"ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778"} Oct 10 15:32:56 crc kubenswrapper[4788]: I1010 15:32:56.220455 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cm877" podStartSLOduration=1.8241636730000002 podStartE2EDuration="3.220432816s" podCreationTimestamp="2025-10-10 15:32:53 +0000 UTC" firstStartedPulling="2025-10-10 15:32:54.171814774 +0000 UTC m=+2876.621530322" lastFinishedPulling="2025-10-10 15:32:55.568083917 +0000 UTC m=+2878.017799465" observedRunningTime="2025-10-10 15:32:56.214658792 +0000 UTC m=+2878.664374340" watchObservedRunningTime="2025-10-10 15:32:56.220432816 +0000 UTC m=+2878.670148374" Oct 10 15:32:58 crc kubenswrapper[4788]: I1010 15:32:58.239099 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:32:58 crc kubenswrapper[4788]: E1010 15:32:58.239364 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:33:03 crc kubenswrapper[4788]: I1010 15:33:03.572107 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:33:03 crc kubenswrapper[4788]: I1010 15:33:03.573229 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:33:03 crc kubenswrapper[4788]: I1010 15:33:03.612718 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:33:04 crc kubenswrapper[4788]: I1010 15:33:04.341061 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:33:04 crc kubenswrapper[4788]: I1010 15:33:04.402907 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm877"] Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.293017 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cm877" podUID="df9de94a-cbdd-48fc-9961-a13710685573" containerName="registry-server" containerID="cri-o://ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778" gracePeriod=2 Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.688338 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.872779 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-utilities\") pod \"df9de94a-cbdd-48fc-9961-a13710685573\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.872873 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfhjt\" (UniqueName: \"kubernetes.io/projected/df9de94a-cbdd-48fc-9961-a13710685573-kube-api-access-nfhjt\") pod \"df9de94a-cbdd-48fc-9961-a13710685573\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.872971 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-catalog-content\") pod \"df9de94a-cbdd-48fc-9961-a13710685573\" (UID: \"df9de94a-cbdd-48fc-9961-a13710685573\") " Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.873986 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-utilities" (OuterVolumeSpecName: "utilities") pod "df9de94a-cbdd-48fc-9961-a13710685573" (UID: "df9de94a-cbdd-48fc-9961-a13710685573"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.878697 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df9de94a-cbdd-48fc-9961-a13710685573-kube-api-access-nfhjt" (OuterVolumeSpecName: "kube-api-access-nfhjt") pod "df9de94a-cbdd-48fc-9961-a13710685573" (UID: "df9de94a-cbdd-48fc-9961-a13710685573"). InnerVolumeSpecName "kube-api-access-nfhjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.888324 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df9de94a-cbdd-48fc-9961-a13710685573" (UID: "df9de94a-cbdd-48fc-9961-a13710685573"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.975076 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.975107 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9de94a-cbdd-48fc-9961-a13710685573-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:33:06 crc kubenswrapper[4788]: I1010 15:33:06.975116 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfhjt\" (UniqueName: \"kubernetes.io/projected/df9de94a-cbdd-48fc-9961-a13710685573-kube-api-access-nfhjt\") on node \"crc\" DevicePath \"\"" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.300708 4788 generic.go:334] "Generic (PLEG): container finished" podID="df9de94a-cbdd-48fc-9961-a13710685573" containerID="ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778" exitCode=0 Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.300808 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cm877" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.300776 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm877" event={"ID":"df9de94a-cbdd-48fc-9961-a13710685573","Type":"ContainerDied","Data":"ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778"} Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.301020 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cm877" event={"ID":"df9de94a-cbdd-48fc-9961-a13710685573","Type":"ContainerDied","Data":"a571642efc39fcc335e5977d503c628c42c5519d81652ca125385ea87d86e586"} Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.301080 4788 scope.go:117] "RemoveContainer" containerID="ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.328457 4788 scope.go:117] "RemoveContainer" containerID="de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.341214 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm877"] Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.351089 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cm877"] Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.359495 4788 scope.go:117] "RemoveContainer" containerID="4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.375374 4788 scope.go:117] "RemoveContainer" containerID="ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778" Oct 10 15:33:07 crc kubenswrapper[4788]: E1010 15:33:07.375815 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778\": container with ID starting with ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778 not found: ID does not exist" containerID="ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.375875 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778"} err="failed to get container status \"ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778\": rpc error: code = NotFound desc = could not find container \"ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778\": container with ID starting with ac2febd266e8b8e08d0b04b26e05ea624250fcca74597be907e137400383f778 not found: ID does not exist" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.375900 4788 scope.go:117] "RemoveContainer" containerID="de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e" Oct 10 15:33:07 crc kubenswrapper[4788]: E1010 15:33:07.376381 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e\": container with ID starting with de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e not found: ID does not exist" containerID="de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.376433 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e"} err="failed to get container status \"de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e\": rpc error: code = NotFound desc = could not find container \"de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e\": container with ID starting with de667b382252f068d8e71e0245c3108505cd443c8118be7d5aa2c1c23f779c3e not found: ID does not exist" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.376450 4788 scope.go:117] "RemoveContainer" containerID="4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81" Oct 10 15:33:07 crc kubenswrapper[4788]: E1010 15:33:07.376792 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81\": container with ID starting with 4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81 not found: ID does not exist" containerID="4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81" Oct 10 15:33:07 crc kubenswrapper[4788]: I1010 15:33:07.376847 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81"} err="failed to get container status \"4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81\": rpc error: code = NotFound desc = could not find container \"4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81\": container with ID starting with 4eaf0f35ddf7ca2180f5f94e3f30db6b9351b2a7d648bbaa0c874d4ac042dc81 not found: ID does not exist" Oct 10 15:33:08 crc kubenswrapper[4788]: I1010 15:33:08.249069 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df9de94a-cbdd-48fc-9961-a13710685573" path="/var/lib/kubelet/pods/df9de94a-cbdd-48fc-9961-a13710685573/volumes" Oct 10 15:33:13 crc kubenswrapper[4788]: I1010 15:33:13.234587 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:33:13 crc kubenswrapper[4788]: E1010 15:33:13.235366 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:33:27 crc kubenswrapper[4788]: I1010 15:33:27.233421 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:33:27 crc kubenswrapper[4788]: E1010 15:33:27.234086 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:33:42 crc kubenswrapper[4788]: I1010 15:33:42.234129 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:33:42 crc kubenswrapper[4788]: E1010 15:33:42.235316 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:33:54 crc kubenswrapper[4788]: I1010 15:33:54.233892 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:33:54 crc kubenswrapper[4788]: E1010 15:33:54.236701 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:34:07 crc kubenswrapper[4788]: I1010 15:34:07.234401 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:34:07 crc kubenswrapper[4788]: E1010 15:34:07.235548 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:34:19 crc kubenswrapper[4788]: I1010 15:34:19.233956 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:34:19 crc kubenswrapper[4788]: E1010 15:34:19.234709 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:34:33 crc kubenswrapper[4788]: I1010 15:34:33.234321 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:34:33 crc kubenswrapper[4788]: E1010 15:34:33.235186 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:34:48 crc kubenswrapper[4788]: I1010 15:34:48.243619 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:34:48 crc kubenswrapper[4788]: E1010 15:34:48.244674 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:35:00 crc kubenswrapper[4788]: I1010 15:35:00.233651 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:35:00 crc kubenswrapper[4788]: E1010 15:35:00.234304 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:35:11 crc kubenswrapper[4788]: I1010 15:35:11.235057 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:35:11 crc kubenswrapper[4788]: E1010 15:35:11.235939 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:35:25 crc kubenswrapper[4788]: I1010 15:35:25.233466 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:35:25 crc kubenswrapper[4788]: E1010 15:35:25.234132 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:35:40 crc kubenswrapper[4788]: I1010 15:35:40.233609 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:35:40 crc kubenswrapper[4788]: E1010 15:35:40.234513 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:35:51 crc kubenswrapper[4788]: I1010 15:35:51.234930 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:35:51 crc kubenswrapper[4788]: E1010 15:35:51.236121 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:36:06 crc kubenswrapper[4788]: I1010 15:36:06.234289 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:36:06 crc kubenswrapper[4788]: E1010 15:36:06.235871 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:36:17 crc kubenswrapper[4788]: I1010 15:36:17.234761 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:36:17 crc kubenswrapper[4788]: E1010 15:36:17.236504 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:36:31 crc kubenswrapper[4788]: I1010 15:36:31.233267 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:36:31 crc kubenswrapper[4788]: E1010 15:36:31.233851 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:36:42 crc kubenswrapper[4788]: I1010 15:36:42.235433 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:36:42 crc kubenswrapper[4788]: E1010 15:36:42.236650 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:36:56 crc kubenswrapper[4788]: I1010 15:36:56.234291 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:36:56 crc kubenswrapper[4788]: E1010 15:36:56.235277 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:37:10 crc kubenswrapper[4788]: I1010 15:37:10.234415 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:37:10 crc kubenswrapper[4788]: E1010 15:37:10.235803 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:37:22 crc kubenswrapper[4788]: I1010 15:37:22.234412 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:37:22 crc kubenswrapper[4788]: E1010 15:37:22.235190 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:37:37 crc kubenswrapper[4788]: I1010 15:37:37.233819 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:37:37 crc kubenswrapper[4788]: I1010 15:37:37.711246 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"9b583e37fb4521523d958e9341d95ad6555f9d867d017e28a5aceaff48b15447"} Oct 10 15:39:18 crc kubenswrapper[4788]: I1010 15:39:18.971281 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gbmtg"] Oct 10 15:39:18 crc kubenswrapper[4788]: E1010 15:39:18.973038 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9de94a-cbdd-48fc-9961-a13710685573" containerName="extract-content" Oct 10 15:39:18 crc kubenswrapper[4788]: I1010 15:39:18.973059 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9de94a-cbdd-48fc-9961-a13710685573" containerName="extract-content" Oct 10 15:39:18 crc kubenswrapper[4788]: E1010 15:39:18.973073 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9de94a-cbdd-48fc-9961-a13710685573" containerName="extract-utilities" Oct 10 15:39:18 crc kubenswrapper[4788]: I1010 15:39:18.973081 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9de94a-cbdd-48fc-9961-a13710685573" containerName="extract-utilities" Oct 10 15:39:18 crc kubenswrapper[4788]: E1010 15:39:18.973090 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9de94a-cbdd-48fc-9961-a13710685573" containerName="registry-server" Oct 10 15:39:18 crc kubenswrapper[4788]: I1010 15:39:18.973097 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9de94a-cbdd-48fc-9961-a13710685573" containerName="registry-server" Oct 10 15:39:18 crc kubenswrapper[4788]: I1010 15:39:18.973326 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="df9de94a-cbdd-48fc-9961-a13710685573" containerName="registry-server" Oct 10 15:39:18 crc kubenswrapper[4788]: I1010 15:39:18.974960 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:18 crc kubenswrapper[4788]: I1010 15:39:18.989908 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gbmtg"] Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.073454 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-utilities\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.073498 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jdkz\" (UniqueName: \"kubernetes.io/projected/cf84f842-f791-41f8-b392-9f4ad72a2867-kube-api-access-6jdkz\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.073602 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-catalog-content\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.174982 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-utilities\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.175037 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jdkz\" (UniqueName: \"kubernetes.io/projected/cf84f842-f791-41f8-b392-9f4ad72a2867-kube-api-access-6jdkz\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.175101 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-catalog-content\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.175565 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-utilities\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.175660 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-catalog-content\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.193964 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jdkz\" (UniqueName: \"kubernetes.io/projected/cf84f842-f791-41f8-b392-9f4ad72a2867-kube-api-access-6jdkz\") pod \"community-operators-gbmtg\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.302236 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:19 crc kubenswrapper[4788]: I1010 15:39:19.806957 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gbmtg"] Oct 10 15:39:20 crc kubenswrapper[4788]: I1010 15:39:20.647484 4788 generic.go:334] "Generic (PLEG): container finished" podID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerID="1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761" exitCode=0 Oct 10 15:39:20 crc kubenswrapper[4788]: I1010 15:39:20.647551 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbmtg" event={"ID":"cf84f842-f791-41f8-b392-9f4ad72a2867","Type":"ContainerDied","Data":"1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761"} Oct 10 15:39:20 crc kubenswrapper[4788]: I1010 15:39:20.647917 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbmtg" event={"ID":"cf84f842-f791-41f8-b392-9f4ad72a2867","Type":"ContainerStarted","Data":"351ef381f5be20ca46a4f0e2cb70298d00748fb30f5af40eb80e5006f025bd59"} Oct 10 15:39:20 crc kubenswrapper[4788]: I1010 15:39:20.649268 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 15:39:21 crc kubenswrapper[4788]: I1010 15:39:21.657677 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbmtg" event={"ID":"cf84f842-f791-41f8-b392-9f4ad72a2867","Type":"ContainerStarted","Data":"cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d"} Oct 10 15:39:22 crc kubenswrapper[4788]: I1010 15:39:22.666813 4788 generic.go:334] "Generic (PLEG): container finished" podID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerID="cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d" exitCode=0 Oct 10 15:39:22 crc kubenswrapper[4788]: I1010 15:39:22.666898 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbmtg" event={"ID":"cf84f842-f791-41f8-b392-9f4ad72a2867","Type":"ContainerDied","Data":"cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d"} Oct 10 15:39:22 crc kubenswrapper[4788]: I1010 15:39:22.667092 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbmtg" event={"ID":"cf84f842-f791-41f8-b392-9f4ad72a2867","Type":"ContainerStarted","Data":"d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87"} Oct 10 15:39:22 crc kubenswrapper[4788]: I1010 15:39:22.684591 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gbmtg" podStartSLOduration=3.220099303 podStartE2EDuration="4.68457113s" podCreationTimestamp="2025-10-10 15:39:18 +0000 UTC" firstStartedPulling="2025-10-10 15:39:20.648958557 +0000 UTC m=+3263.098674105" lastFinishedPulling="2025-10-10 15:39:22.113430384 +0000 UTC m=+3264.563145932" observedRunningTime="2025-10-10 15:39:22.680869602 +0000 UTC m=+3265.130585170" watchObservedRunningTime="2025-10-10 15:39:22.68457113 +0000 UTC m=+3265.134286678" Oct 10 15:39:29 crc kubenswrapper[4788]: I1010 15:39:29.303394 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:29 crc kubenswrapper[4788]: I1010 15:39:29.303983 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:29 crc kubenswrapper[4788]: I1010 15:39:29.383089 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:29 crc kubenswrapper[4788]: I1010 15:39:29.780221 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:29 crc kubenswrapper[4788]: I1010 15:39:29.829910 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gbmtg"] Oct 10 15:39:31 crc kubenswrapper[4788]: I1010 15:39:31.739505 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gbmtg" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerName="registry-server" containerID="cri-o://d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87" gracePeriod=2 Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.621360 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.695071 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-utilities\") pod \"cf84f842-f791-41f8-b392-9f4ad72a2867\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.695240 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jdkz\" (UniqueName: \"kubernetes.io/projected/cf84f842-f791-41f8-b392-9f4ad72a2867-kube-api-access-6jdkz\") pod \"cf84f842-f791-41f8-b392-9f4ad72a2867\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.695353 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-catalog-content\") pod \"cf84f842-f791-41f8-b392-9f4ad72a2867\" (UID: \"cf84f842-f791-41f8-b392-9f4ad72a2867\") " Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.695886 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-utilities" (OuterVolumeSpecName: "utilities") pod "cf84f842-f791-41f8-b392-9f4ad72a2867" (UID: "cf84f842-f791-41f8-b392-9f4ad72a2867"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.700558 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf84f842-f791-41f8-b392-9f4ad72a2867-kube-api-access-6jdkz" (OuterVolumeSpecName: "kube-api-access-6jdkz") pod "cf84f842-f791-41f8-b392-9f4ad72a2867" (UID: "cf84f842-f791-41f8-b392-9f4ad72a2867"). InnerVolumeSpecName "kube-api-access-6jdkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.748057 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf84f842-f791-41f8-b392-9f4ad72a2867" (UID: "cf84f842-f791-41f8-b392-9f4ad72a2867"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.749472 4788 generic.go:334] "Generic (PLEG): container finished" podID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerID="d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87" exitCode=0 Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.749518 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbmtg" event={"ID":"cf84f842-f791-41f8-b392-9f4ad72a2867","Type":"ContainerDied","Data":"d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87"} Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.749556 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gbmtg" event={"ID":"cf84f842-f791-41f8-b392-9f4ad72a2867","Type":"ContainerDied","Data":"351ef381f5be20ca46a4f0e2cb70298d00748fb30f5af40eb80e5006f025bd59"} Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.749582 4788 scope.go:117] "RemoveContainer" containerID="d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.749709 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gbmtg" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.765964 4788 scope.go:117] "RemoveContainer" containerID="cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.782612 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gbmtg"] Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.787559 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gbmtg"] Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.791750 4788 scope.go:117] "RemoveContainer" containerID="1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.796855 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.796885 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jdkz\" (UniqueName: \"kubernetes.io/projected/cf84f842-f791-41f8-b392-9f4ad72a2867-kube-api-access-6jdkz\") on node \"crc\" DevicePath \"\"" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.796900 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf84f842-f791-41f8-b392-9f4ad72a2867-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.822084 4788 scope.go:117] "RemoveContainer" containerID="d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87" Oct 10 15:39:32 crc kubenswrapper[4788]: E1010 15:39:32.822598 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87\": container with ID starting with d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87 not found: ID does not exist" containerID="d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.822706 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87"} err="failed to get container status \"d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87\": rpc error: code = NotFound desc = could not find container \"d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87\": container with ID starting with d9f5f0afb8e62c374a1b730978eb964f7cff1a5d765cd2869415813d15b2fa87 not found: ID does not exist" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.822812 4788 scope.go:117] "RemoveContainer" containerID="cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d" Oct 10 15:39:32 crc kubenswrapper[4788]: E1010 15:39:32.823259 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d\": container with ID starting with cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d not found: ID does not exist" containerID="cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.823292 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d"} err="failed to get container status \"cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d\": rpc error: code = NotFound desc = could not find container \"cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d\": container with ID starting with cb94bd7f5f4ffd0769afb6434908c6175a248de2967cd7bd54c0f0575f03eb7d not found: ID does not exist" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.823314 4788 scope.go:117] "RemoveContainer" containerID="1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761" Oct 10 15:39:32 crc kubenswrapper[4788]: E1010 15:39:32.823502 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761\": container with ID starting with 1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761 not found: ID does not exist" containerID="1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761" Oct 10 15:39:32 crc kubenswrapper[4788]: I1010 15:39:32.823523 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761"} err="failed to get container status \"1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761\": rpc error: code = NotFound desc = could not find container \"1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761\": container with ID starting with 1fc3b2d5612a4dbb0f696bffaee46e4e2be378b1dc23cfac5f6f78147592d761 not found: ID does not exist" Oct 10 15:39:34 crc kubenswrapper[4788]: I1010 15:39:34.249827 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" path="/var/lib/kubelet/pods/cf84f842-f791-41f8-b392-9f4ad72a2867/volumes" Oct 10 15:39:59 crc kubenswrapper[4788]: I1010 15:39:59.406056 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:39:59 crc kubenswrapper[4788]: I1010 15:39:59.406870 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.850057 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dmt9z"] Oct 10 15:40:15 crc kubenswrapper[4788]: E1010 15:40:15.851280 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerName="registry-server" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.851305 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerName="registry-server" Oct 10 15:40:15 crc kubenswrapper[4788]: E1010 15:40:15.851343 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerName="extract-content" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.851355 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerName="extract-content" Oct 10 15:40:15 crc kubenswrapper[4788]: E1010 15:40:15.851381 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerName="extract-utilities" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.851394 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerName="extract-utilities" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.851729 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf84f842-f791-41f8-b392-9f4ad72a2867" containerName="registry-server" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.855190 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.875569 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dmt9z"] Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.983815 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-catalog-content\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.984317 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-utilities\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:15 crc kubenswrapper[4788]: I1010 15:40:15.984415 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qfzb\" (UniqueName: \"kubernetes.io/projected/778bb6a8-049c-4b47-a8bc-b6138711ab59-kube-api-access-6qfzb\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:16 crc kubenswrapper[4788]: I1010 15:40:16.085576 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-utilities\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:16 crc kubenswrapper[4788]: I1010 15:40:16.085645 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qfzb\" (UniqueName: \"kubernetes.io/projected/778bb6a8-049c-4b47-a8bc-b6138711ab59-kube-api-access-6qfzb\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:16 crc kubenswrapper[4788]: I1010 15:40:16.085719 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-catalog-content\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:16 crc kubenswrapper[4788]: I1010 15:40:16.086268 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-catalog-content\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:16 crc kubenswrapper[4788]: I1010 15:40:16.086269 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-utilities\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:16 crc kubenswrapper[4788]: I1010 15:40:16.107106 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qfzb\" (UniqueName: \"kubernetes.io/projected/778bb6a8-049c-4b47-a8bc-b6138711ab59-kube-api-access-6qfzb\") pod \"redhat-operators-dmt9z\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:16 crc kubenswrapper[4788]: I1010 15:40:16.227792 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:16 crc kubenswrapper[4788]: I1010 15:40:16.671110 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dmt9z"] Oct 10 15:40:17 crc kubenswrapper[4788]: I1010 15:40:17.134474 4788 generic.go:334] "Generic (PLEG): container finished" podID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerID="ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3" exitCode=0 Oct 10 15:40:17 crc kubenswrapper[4788]: I1010 15:40:17.134552 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmt9z" event={"ID":"778bb6a8-049c-4b47-a8bc-b6138711ab59","Type":"ContainerDied","Data":"ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3"} Oct 10 15:40:17 crc kubenswrapper[4788]: I1010 15:40:17.134612 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmt9z" event={"ID":"778bb6a8-049c-4b47-a8bc-b6138711ab59","Type":"ContainerStarted","Data":"bdd7bf79aea4b22cfc1a236a8dce5b7912ebfbffc73764ff42576a10e4601e32"} Oct 10 15:40:18 crc kubenswrapper[4788]: I1010 15:40:18.143299 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmt9z" event={"ID":"778bb6a8-049c-4b47-a8bc-b6138711ab59","Type":"ContainerStarted","Data":"7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e"} Oct 10 15:40:19 crc kubenswrapper[4788]: I1010 15:40:19.156879 4788 generic.go:334] "Generic (PLEG): container finished" podID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerID="7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e" exitCode=0 Oct 10 15:40:19 crc kubenswrapper[4788]: I1010 15:40:19.156942 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmt9z" event={"ID":"778bb6a8-049c-4b47-a8bc-b6138711ab59","Type":"ContainerDied","Data":"7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e"} Oct 10 15:40:20 crc kubenswrapper[4788]: I1010 15:40:20.167379 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmt9z" event={"ID":"778bb6a8-049c-4b47-a8bc-b6138711ab59","Type":"ContainerStarted","Data":"1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098"} Oct 10 15:40:20 crc kubenswrapper[4788]: I1010 15:40:20.189664 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dmt9z" podStartSLOduration=2.778379004 podStartE2EDuration="5.189645132s" podCreationTimestamp="2025-10-10 15:40:15 +0000 UTC" firstStartedPulling="2025-10-10 15:40:17.137410534 +0000 UTC m=+3319.587126082" lastFinishedPulling="2025-10-10 15:40:19.548676672 +0000 UTC m=+3321.998392210" observedRunningTime="2025-10-10 15:40:20.184688091 +0000 UTC m=+3322.634403639" watchObservedRunningTime="2025-10-10 15:40:20.189645132 +0000 UTC m=+3322.639360670" Oct 10 15:40:26 crc kubenswrapper[4788]: I1010 15:40:26.228408 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:26 crc kubenswrapper[4788]: I1010 15:40:26.229837 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:26 crc kubenswrapper[4788]: I1010 15:40:26.271847 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:27 crc kubenswrapper[4788]: I1010 15:40:27.262689 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:27 crc kubenswrapper[4788]: I1010 15:40:27.326296 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dmt9z"] Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.233748 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dmt9z" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerName="registry-server" containerID="cri-o://1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098" gracePeriod=2 Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.407261 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.407330 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.697780 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.889790 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qfzb\" (UniqueName: \"kubernetes.io/projected/778bb6a8-049c-4b47-a8bc-b6138711ab59-kube-api-access-6qfzb\") pod \"778bb6a8-049c-4b47-a8bc-b6138711ab59\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.890274 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-catalog-content\") pod \"778bb6a8-049c-4b47-a8bc-b6138711ab59\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.890478 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-utilities\") pod \"778bb6a8-049c-4b47-a8bc-b6138711ab59\" (UID: \"778bb6a8-049c-4b47-a8bc-b6138711ab59\") " Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.891650 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-utilities" (OuterVolumeSpecName: "utilities") pod "778bb6a8-049c-4b47-a8bc-b6138711ab59" (UID: "778bb6a8-049c-4b47-a8bc-b6138711ab59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.898647 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778bb6a8-049c-4b47-a8bc-b6138711ab59-kube-api-access-6qfzb" (OuterVolumeSpecName: "kube-api-access-6qfzb") pod "778bb6a8-049c-4b47-a8bc-b6138711ab59" (UID: "778bb6a8-049c-4b47-a8bc-b6138711ab59"). InnerVolumeSpecName "kube-api-access-6qfzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.992443 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qfzb\" (UniqueName: \"kubernetes.io/projected/778bb6a8-049c-4b47-a8bc-b6138711ab59-kube-api-access-6qfzb\") on node \"crc\" DevicePath \"\"" Oct 10 15:40:29 crc kubenswrapper[4788]: I1010 15:40:29.992496 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.041608 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "778bb6a8-049c-4b47-a8bc-b6138711ab59" (UID: "778bb6a8-049c-4b47-a8bc-b6138711ab59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.094111 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/778bb6a8-049c-4b47-a8bc-b6138711ab59-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.256764 4788 generic.go:334] "Generic (PLEG): container finished" podID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerID="1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098" exitCode=0 Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.256827 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmt9z" event={"ID":"778bb6a8-049c-4b47-a8bc-b6138711ab59","Type":"ContainerDied","Data":"1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098"} Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.256869 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmt9z" event={"ID":"778bb6a8-049c-4b47-a8bc-b6138711ab59","Type":"ContainerDied","Data":"bdd7bf79aea4b22cfc1a236a8dce5b7912ebfbffc73764ff42576a10e4601e32"} Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.256887 4788 scope.go:117] "RemoveContainer" containerID="1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.256940 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmt9z" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.317474 4788 scope.go:117] "RemoveContainer" containerID="7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.319751 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dmt9z"] Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.334586 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dmt9z"] Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.343092 4788 scope.go:117] "RemoveContainer" containerID="ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.365464 4788 scope.go:117] "RemoveContainer" containerID="1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098" Oct 10 15:40:30 crc kubenswrapper[4788]: E1010 15:40:30.366272 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098\": container with ID starting with 1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098 not found: ID does not exist" containerID="1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.366341 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098"} err="failed to get container status \"1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098\": rpc error: code = NotFound desc = could not find container \"1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098\": container with ID starting with 1ec3af7a0b2e85f3997dae38aecba7cb44511588c8221ee43c960b91d84d1098 not found: ID does not exist" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.366403 4788 scope.go:117] "RemoveContainer" containerID="7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e" Oct 10 15:40:30 crc kubenswrapper[4788]: E1010 15:40:30.366958 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e\": container with ID starting with 7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e not found: ID does not exist" containerID="7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.367016 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e"} err="failed to get container status \"7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e\": rpc error: code = NotFound desc = could not find container \"7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e\": container with ID starting with 7aa8d9265487582045c27c05aef2bf83a3a775f9c95f91c5c040733255ae396e not found: ID does not exist" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.367055 4788 scope.go:117] "RemoveContainer" containerID="ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3" Oct 10 15:40:30 crc kubenswrapper[4788]: E1010 15:40:30.367496 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3\": container with ID starting with ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3 not found: ID does not exist" containerID="ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3" Oct 10 15:40:30 crc kubenswrapper[4788]: I1010 15:40:30.367554 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3"} err="failed to get container status \"ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3\": rpc error: code = NotFound desc = could not find container \"ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3\": container with ID starting with ea1e78849eb9ff053dad0753b1e32c30f589b72085bbf520ed16d775094466e3 not found: ID does not exist" Oct 10 15:40:32 crc kubenswrapper[4788]: I1010 15:40:32.242928 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" path="/var/lib/kubelet/pods/778bb6a8-049c-4b47-a8bc-b6138711ab59/volumes" Oct 10 15:40:59 crc kubenswrapper[4788]: I1010 15:40:59.406671 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:40:59 crc kubenswrapper[4788]: I1010 15:40:59.408521 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:40:59 crc kubenswrapper[4788]: I1010 15:40:59.408728 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:40:59 crc kubenswrapper[4788]: I1010 15:40:59.409626 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b583e37fb4521523d958e9341d95ad6555f9d867d017e28a5aceaff48b15447"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:40:59 crc kubenswrapper[4788]: I1010 15:40:59.409807 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://9b583e37fb4521523d958e9341d95ad6555f9d867d017e28a5aceaff48b15447" gracePeriod=600 Oct 10 15:41:00 crc kubenswrapper[4788]: I1010 15:41:00.504450 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="9b583e37fb4521523d958e9341d95ad6555f9d867d017e28a5aceaff48b15447" exitCode=0 Oct 10 15:41:00 crc kubenswrapper[4788]: I1010 15:41:00.504556 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"9b583e37fb4521523d958e9341d95ad6555f9d867d017e28a5aceaff48b15447"} Oct 10 15:41:00 crc kubenswrapper[4788]: I1010 15:41:00.505893 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5"} Oct 10 15:41:00 crc kubenswrapper[4788]: I1010 15:41:00.505947 4788 scope.go:117] "RemoveContainer" containerID="42e98650cc5692cdcc74c6bf39450f7e19abacbaa0c2e08a418d51bd9d59ab51" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.815083 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5pfrh"] Oct 10 15:41:03 crc kubenswrapper[4788]: E1010 15:41:03.816030 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerName="extract-content" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.816085 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerName="extract-content" Oct 10 15:41:03 crc kubenswrapper[4788]: E1010 15:41:03.816121 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerName="registry-server" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.816133 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerName="registry-server" Oct 10 15:41:03 crc kubenswrapper[4788]: E1010 15:41:03.816190 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerName="extract-utilities" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.816203 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerName="extract-utilities" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.816463 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="778bb6a8-049c-4b47-a8bc-b6138711ab59" containerName="registry-server" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.817856 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.842666 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5pfrh"] Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.914872 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-utilities\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.914973 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-catalog-content\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:03 crc kubenswrapper[4788]: I1010 15:41:03.915012 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcdps\" (UniqueName: \"kubernetes.io/projected/88438649-52d8-462b-9a8b-f2cab293d730-kube-api-access-fcdps\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:04 crc kubenswrapper[4788]: I1010 15:41:04.016081 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcdps\" (UniqueName: \"kubernetes.io/projected/88438649-52d8-462b-9a8b-f2cab293d730-kube-api-access-fcdps\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:04 crc kubenswrapper[4788]: I1010 15:41:04.016206 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-utilities\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:04 crc kubenswrapper[4788]: I1010 15:41:04.016276 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-catalog-content\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:04 crc kubenswrapper[4788]: I1010 15:41:04.016845 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-catalog-content\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:04 crc kubenswrapper[4788]: I1010 15:41:04.017056 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-utilities\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:04 crc kubenswrapper[4788]: I1010 15:41:04.046709 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcdps\" (UniqueName: \"kubernetes.io/projected/88438649-52d8-462b-9a8b-f2cab293d730-kube-api-access-fcdps\") pod \"certified-operators-5pfrh\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:04 crc kubenswrapper[4788]: I1010 15:41:04.160293 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:04 crc kubenswrapper[4788]: I1010 15:41:04.639789 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5pfrh"] Oct 10 15:41:05 crc kubenswrapper[4788]: I1010 15:41:05.553736 4788 generic.go:334] "Generic (PLEG): container finished" podID="88438649-52d8-462b-9a8b-f2cab293d730" containerID="f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798" exitCode=0 Oct 10 15:41:05 crc kubenswrapper[4788]: I1010 15:41:05.553850 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5pfrh" event={"ID":"88438649-52d8-462b-9a8b-f2cab293d730","Type":"ContainerDied","Data":"f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798"} Oct 10 15:41:05 crc kubenswrapper[4788]: I1010 15:41:05.554321 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5pfrh" event={"ID":"88438649-52d8-462b-9a8b-f2cab293d730","Type":"ContainerStarted","Data":"84cc1d373947ee00b9a008a19fac99452da3f01491fb8265fb1a972095041b19"} Oct 10 15:41:06 crc kubenswrapper[4788]: I1010 15:41:06.566007 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5pfrh" event={"ID":"88438649-52d8-462b-9a8b-f2cab293d730","Type":"ContainerStarted","Data":"223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b"} Oct 10 15:41:07 crc kubenswrapper[4788]: I1010 15:41:07.576529 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5pfrh" event={"ID":"88438649-52d8-462b-9a8b-f2cab293d730","Type":"ContainerDied","Data":"223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b"} Oct 10 15:41:07 crc kubenswrapper[4788]: I1010 15:41:07.576611 4788 generic.go:334] "Generic (PLEG): container finished" podID="88438649-52d8-462b-9a8b-f2cab293d730" containerID="223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b" exitCode=0 Oct 10 15:41:08 crc kubenswrapper[4788]: I1010 15:41:08.585545 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5pfrh" event={"ID":"88438649-52d8-462b-9a8b-f2cab293d730","Type":"ContainerStarted","Data":"77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6"} Oct 10 15:41:08 crc kubenswrapper[4788]: I1010 15:41:08.609442 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5pfrh" podStartSLOduration=3.072455553 podStartE2EDuration="5.609419019s" podCreationTimestamp="2025-10-10 15:41:03 +0000 UTC" firstStartedPulling="2025-10-10 15:41:05.556559734 +0000 UTC m=+3368.006275292" lastFinishedPulling="2025-10-10 15:41:08.0935232 +0000 UTC m=+3370.543238758" observedRunningTime="2025-10-10 15:41:08.604334604 +0000 UTC m=+3371.054050152" watchObservedRunningTime="2025-10-10 15:41:08.609419019 +0000 UTC m=+3371.059134567" Oct 10 15:41:14 crc kubenswrapper[4788]: I1010 15:41:14.160921 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:14 crc kubenswrapper[4788]: I1010 15:41:14.161243 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:14 crc kubenswrapper[4788]: I1010 15:41:14.240704 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:14 crc kubenswrapper[4788]: I1010 15:41:14.708238 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:14 crc kubenswrapper[4788]: I1010 15:41:14.764080 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5pfrh"] Oct 10 15:41:16 crc kubenswrapper[4788]: I1010 15:41:16.678549 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5pfrh" podUID="88438649-52d8-462b-9a8b-f2cab293d730" containerName="registry-server" containerID="cri-o://77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6" gracePeriod=2 Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.184366 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.338657 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-catalog-content\") pod \"88438649-52d8-462b-9a8b-f2cab293d730\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.339003 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcdps\" (UniqueName: \"kubernetes.io/projected/88438649-52d8-462b-9a8b-f2cab293d730-kube-api-access-fcdps\") pod \"88438649-52d8-462b-9a8b-f2cab293d730\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.339034 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-utilities\") pod \"88438649-52d8-462b-9a8b-f2cab293d730\" (UID: \"88438649-52d8-462b-9a8b-f2cab293d730\") " Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.341132 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-utilities" (OuterVolumeSpecName: "utilities") pod "88438649-52d8-462b-9a8b-f2cab293d730" (UID: "88438649-52d8-462b-9a8b-f2cab293d730"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.347161 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88438649-52d8-462b-9a8b-f2cab293d730-kube-api-access-fcdps" (OuterVolumeSpecName: "kube-api-access-fcdps") pod "88438649-52d8-462b-9a8b-f2cab293d730" (UID: "88438649-52d8-462b-9a8b-f2cab293d730"). InnerVolumeSpecName "kube-api-access-fcdps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.393106 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88438649-52d8-462b-9a8b-f2cab293d730" (UID: "88438649-52d8-462b-9a8b-f2cab293d730"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.440942 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcdps\" (UniqueName: \"kubernetes.io/projected/88438649-52d8-462b-9a8b-f2cab293d730-kube-api-access-fcdps\") on node \"crc\" DevicePath \"\"" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.440987 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.440999 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88438649-52d8-462b-9a8b-f2cab293d730-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.689386 4788 generic.go:334] "Generic (PLEG): container finished" podID="88438649-52d8-462b-9a8b-f2cab293d730" containerID="77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6" exitCode=0 Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.689430 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5pfrh" event={"ID":"88438649-52d8-462b-9a8b-f2cab293d730","Type":"ContainerDied","Data":"77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6"} Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.689476 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5pfrh" event={"ID":"88438649-52d8-462b-9a8b-f2cab293d730","Type":"ContainerDied","Data":"84cc1d373947ee00b9a008a19fac99452da3f01491fb8265fb1a972095041b19"} Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.689510 4788 scope.go:117] "RemoveContainer" containerID="77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.689523 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5pfrh" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.743132 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5pfrh"] Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.745658 4788 scope.go:117] "RemoveContainer" containerID="223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.755384 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5pfrh"] Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.781121 4788 scope.go:117] "RemoveContainer" containerID="f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.801923 4788 scope.go:117] "RemoveContainer" containerID="77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6" Oct 10 15:41:17 crc kubenswrapper[4788]: E1010 15:41:17.802378 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6\": container with ID starting with 77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6 not found: ID does not exist" containerID="77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.802509 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6"} err="failed to get container status \"77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6\": rpc error: code = NotFound desc = could not find container \"77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6\": container with ID starting with 77f5dfebc8d71ce5aa218d75418f20f60467250531ed552dc20fa3dbd2edbba6 not found: ID does not exist" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.802601 4788 scope.go:117] "RemoveContainer" containerID="223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b" Oct 10 15:41:17 crc kubenswrapper[4788]: E1010 15:41:17.803011 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b\": container with ID starting with 223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b not found: ID does not exist" containerID="223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.803033 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b"} err="failed to get container status \"223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b\": rpc error: code = NotFound desc = could not find container \"223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b\": container with ID starting with 223acacfb2706802c60e2642c3c6c3e8d6db120212885f14ad29bcba0e623a4b not found: ID does not exist" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.803053 4788 scope.go:117] "RemoveContainer" containerID="f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798" Oct 10 15:41:17 crc kubenswrapper[4788]: E1010 15:41:17.803319 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798\": container with ID starting with f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798 not found: ID does not exist" containerID="f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798" Oct 10 15:41:17 crc kubenswrapper[4788]: I1010 15:41:17.803393 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798"} err="failed to get container status \"f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798\": rpc error: code = NotFound desc = could not find container \"f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798\": container with ID starting with f22a6131c91f983feb82b94d8dd25bbee26efb6a4d5c03817882111960e45798 not found: ID does not exist" Oct 10 15:41:18 crc kubenswrapper[4788]: I1010 15:41:18.249869 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88438649-52d8-462b-9a8b-f2cab293d730" path="/var/lib/kubelet/pods/88438649-52d8-462b-9a8b-f2cab293d730/volumes" Oct 10 15:42:59 crc kubenswrapper[4788]: I1010 15:42:59.406533 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:42:59 crc kubenswrapper[4788]: I1010 15:42:59.407198 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:43:29 crc kubenswrapper[4788]: I1010 15:43:29.406369 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:43:29 crc kubenswrapper[4788]: I1010 15:43:29.406760 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:43:59 crc kubenswrapper[4788]: I1010 15:43:59.406543 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:43:59 crc kubenswrapper[4788]: I1010 15:43:59.407477 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:43:59 crc kubenswrapper[4788]: I1010 15:43:59.407566 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:43:59 crc kubenswrapper[4788]: I1010 15:43:59.408743 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:43:59 crc kubenswrapper[4788]: I1010 15:43:59.408853 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" gracePeriod=600 Oct 10 15:43:59 crc kubenswrapper[4788]: E1010 15:43:59.532534 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:44:00 crc kubenswrapper[4788]: I1010 15:44:00.224684 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" exitCode=0 Oct 10 15:44:00 crc kubenswrapper[4788]: I1010 15:44:00.224740 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5"} Oct 10 15:44:00 crc kubenswrapper[4788]: I1010 15:44:00.225256 4788 scope.go:117] "RemoveContainer" containerID="9b583e37fb4521523d958e9341d95ad6555f9d867d017e28a5aceaff48b15447" Oct 10 15:44:00 crc kubenswrapper[4788]: I1010 15:44:00.226558 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:44:00 crc kubenswrapper[4788]: E1010 15:44:00.227117 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.754551 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-clbfh"] Oct 10 15:44:09 crc kubenswrapper[4788]: E1010 15:44:09.755575 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88438649-52d8-462b-9a8b-f2cab293d730" containerName="extract-content" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.755597 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="88438649-52d8-462b-9a8b-f2cab293d730" containerName="extract-content" Oct 10 15:44:09 crc kubenswrapper[4788]: E1010 15:44:09.755621 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88438649-52d8-462b-9a8b-f2cab293d730" containerName="extract-utilities" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.755632 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="88438649-52d8-462b-9a8b-f2cab293d730" containerName="extract-utilities" Oct 10 15:44:09 crc kubenswrapper[4788]: E1010 15:44:09.755667 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88438649-52d8-462b-9a8b-f2cab293d730" containerName="registry-server" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.755679 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="88438649-52d8-462b-9a8b-f2cab293d730" containerName="registry-server" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.755899 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="88438649-52d8-462b-9a8b-f2cab293d730" containerName="registry-server" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.757809 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.763118 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-clbfh"] Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.940305 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-catalog-content\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.940363 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8r4v\" (UniqueName: \"kubernetes.io/projected/b33aacfd-c55e-47a5-93cd-e426bba74235-kube-api-access-d8r4v\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:09 crc kubenswrapper[4788]: I1010 15:44:09.940534 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-utilities\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:10 crc kubenswrapper[4788]: I1010 15:44:10.042612 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-catalog-content\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:10 crc kubenswrapper[4788]: I1010 15:44:10.042680 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8r4v\" (UniqueName: \"kubernetes.io/projected/b33aacfd-c55e-47a5-93cd-e426bba74235-kube-api-access-d8r4v\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:10 crc kubenswrapper[4788]: I1010 15:44:10.042739 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-utilities\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:10 crc kubenswrapper[4788]: I1010 15:44:10.043131 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-catalog-content\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:10 crc kubenswrapper[4788]: I1010 15:44:10.043334 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-utilities\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:10 crc kubenswrapper[4788]: I1010 15:44:10.068978 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8r4v\" (UniqueName: \"kubernetes.io/projected/b33aacfd-c55e-47a5-93cd-e426bba74235-kube-api-access-d8r4v\") pod \"redhat-marketplace-clbfh\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:10 crc kubenswrapper[4788]: I1010 15:44:10.081870 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:10 crc kubenswrapper[4788]: I1010 15:44:10.642865 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-clbfh"] Oct 10 15:44:11 crc kubenswrapper[4788]: I1010 15:44:11.325662 4788 generic.go:334] "Generic (PLEG): container finished" podID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerID="02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83" exitCode=0 Oct 10 15:44:11 crc kubenswrapper[4788]: I1010 15:44:11.325728 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clbfh" event={"ID":"b33aacfd-c55e-47a5-93cd-e426bba74235","Type":"ContainerDied","Data":"02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83"} Oct 10 15:44:11 crc kubenswrapper[4788]: I1010 15:44:11.325769 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clbfh" event={"ID":"b33aacfd-c55e-47a5-93cd-e426bba74235","Type":"ContainerStarted","Data":"24955aa1007830615dba1a67e18451c2f2ed79cae4eac1b4cbcfc749bfc5d7df"} Oct 10 15:44:12 crc kubenswrapper[4788]: I1010 15:44:12.339376 4788 generic.go:334] "Generic (PLEG): container finished" podID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerID="5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776" exitCode=0 Oct 10 15:44:12 crc kubenswrapper[4788]: I1010 15:44:12.339472 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clbfh" event={"ID":"b33aacfd-c55e-47a5-93cd-e426bba74235","Type":"ContainerDied","Data":"5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776"} Oct 10 15:44:13 crc kubenswrapper[4788]: I1010 15:44:13.233212 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:44:13 crc kubenswrapper[4788]: E1010 15:44:13.233696 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:44:13 crc kubenswrapper[4788]: I1010 15:44:13.350986 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clbfh" event={"ID":"b33aacfd-c55e-47a5-93cd-e426bba74235","Type":"ContainerStarted","Data":"fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100"} Oct 10 15:44:13 crc kubenswrapper[4788]: I1010 15:44:13.373732 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-clbfh" podStartSLOduration=2.871941596 podStartE2EDuration="4.37371345s" podCreationTimestamp="2025-10-10 15:44:09 +0000 UTC" firstStartedPulling="2025-10-10 15:44:11.331716982 +0000 UTC m=+3553.781432540" lastFinishedPulling="2025-10-10 15:44:12.833488846 +0000 UTC m=+3555.283204394" observedRunningTime="2025-10-10 15:44:13.368564205 +0000 UTC m=+3555.818279763" watchObservedRunningTime="2025-10-10 15:44:13.37371345 +0000 UTC m=+3555.823429008" Oct 10 15:44:20 crc kubenswrapper[4788]: I1010 15:44:20.082422 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:20 crc kubenswrapper[4788]: I1010 15:44:20.083110 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:20 crc kubenswrapper[4788]: I1010 15:44:20.156925 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:20 crc kubenswrapper[4788]: I1010 15:44:20.474053 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:22 crc kubenswrapper[4788]: I1010 15:44:22.727498 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-clbfh"] Oct 10 15:44:22 crc kubenswrapper[4788]: I1010 15:44:22.728284 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-clbfh" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerName="registry-server" containerID="cri-o://fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100" gracePeriod=2 Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.117506 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.167360 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8r4v\" (UniqueName: \"kubernetes.io/projected/b33aacfd-c55e-47a5-93cd-e426bba74235-kube-api-access-d8r4v\") pod \"b33aacfd-c55e-47a5-93cd-e426bba74235\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.167443 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-utilities\") pod \"b33aacfd-c55e-47a5-93cd-e426bba74235\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.167521 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-catalog-content\") pod \"b33aacfd-c55e-47a5-93cd-e426bba74235\" (UID: \"b33aacfd-c55e-47a5-93cd-e426bba74235\") " Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.168477 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-utilities" (OuterVolumeSpecName: "utilities") pod "b33aacfd-c55e-47a5-93cd-e426bba74235" (UID: "b33aacfd-c55e-47a5-93cd-e426bba74235"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.176612 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b33aacfd-c55e-47a5-93cd-e426bba74235-kube-api-access-d8r4v" (OuterVolumeSpecName: "kube-api-access-d8r4v") pod "b33aacfd-c55e-47a5-93cd-e426bba74235" (UID: "b33aacfd-c55e-47a5-93cd-e426bba74235"). InnerVolumeSpecName "kube-api-access-d8r4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.184286 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b33aacfd-c55e-47a5-93cd-e426bba74235" (UID: "b33aacfd-c55e-47a5-93cd-e426bba74235"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.269312 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8r4v\" (UniqueName: \"kubernetes.io/projected/b33aacfd-c55e-47a5-93cd-e426bba74235-kube-api-access-d8r4v\") on node \"crc\" DevicePath \"\"" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.269352 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.269365 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b33aacfd-c55e-47a5-93cd-e426bba74235-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.453551 4788 generic.go:334] "Generic (PLEG): container finished" podID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerID="fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100" exitCode=0 Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.453595 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clbfh" event={"ID":"b33aacfd-c55e-47a5-93cd-e426bba74235","Type":"ContainerDied","Data":"fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100"} Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.453622 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-clbfh" event={"ID":"b33aacfd-c55e-47a5-93cd-e426bba74235","Type":"ContainerDied","Data":"24955aa1007830615dba1a67e18451c2f2ed79cae4eac1b4cbcfc749bfc5d7df"} Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.453638 4788 scope.go:117] "RemoveContainer" containerID="fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.453678 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-clbfh" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.484983 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-clbfh"] Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.489473 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-clbfh"] Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.497542 4788 scope.go:117] "RemoveContainer" containerID="5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.529809 4788 scope.go:117] "RemoveContainer" containerID="02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.551601 4788 scope.go:117] "RemoveContainer" containerID="fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100" Oct 10 15:44:23 crc kubenswrapper[4788]: E1010 15:44:23.552295 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100\": container with ID starting with fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100 not found: ID does not exist" containerID="fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.552342 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100"} err="failed to get container status \"fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100\": rpc error: code = NotFound desc = could not find container \"fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100\": container with ID starting with fb448c1d686a37600eab613785c45498ba55f02f97614319193125e3240b3100 not found: ID does not exist" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.552368 4788 scope.go:117] "RemoveContainer" containerID="5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776" Oct 10 15:44:23 crc kubenswrapper[4788]: E1010 15:44:23.552786 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776\": container with ID starting with 5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776 not found: ID does not exist" containerID="5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.552808 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776"} err="failed to get container status \"5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776\": rpc error: code = NotFound desc = could not find container \"5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776\": container with ID starting with 5349a508b3b2d5fd92514497a94e8284749c3d3de2054fa79d72bafdf2bcc776 not found: ID does not exist" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.552824 4788 scope.go:117] "RemoveContainer" containerID="02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83" Oct 10 15:44:23 crc kubenswrapper[4788]: E1010 15:44:23.554515 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83\": container with ID starting with 02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83 not found: ID does not exist" containerID="02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83" Oct 10 15:44:23 crc kubenswrapper[4788]: I1010 15:44:23.554541 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83"} err="failed to get container status \"02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83\": rpc error: code = NotFound desc = could not find container \"02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83\": container with ID starting with 02e0fcfd3a7632a54d7795fe65bb43b90d9e6d03780bf1a78bd298bfb16b6f83 not found: ID does not exist" Oct 10 15:44:24 crc kubenswrapper[4788]: I1010 15:44:24.251057 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" path="/var/lib/kubelet/pods/b33aacfd-c55e-47a5-93cd-e426bba74235/volumes" Oct 10 15:44:25 crc kubenswrapper[4788]: I1010 15:44:25.234467 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:44:25 crc kubenswrapper[4788]: E1010 15:44:25.235004 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:44:38 crc kubenswrapper[4788]: I1010 15:44:38.237003 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:44:38 crc kubenswrapper[4788]: E1010 15:44:38.237717 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:44:50 crc kubenswrapper[4788]: I1010 15:44:50.234099 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:44:50 crc kubenswrapper[4788]: E1010 15:44:50.235071 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.203782 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd"] Oct 10 15:45:00 crc kubenswrapper[4788]: E1010 15:45:00.204748 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerName="extract-content" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.204766 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerName="extract-content" Oct 10 15:45:00 crc kubenswrapper[4788]: E1010 15:45:00.204783 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerName="registry-server" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.204792 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerName="registry-server" Oct 10 15:45:00 crc kubenswrapper[4788]: E1010 15:45:00.204810 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerName="extract-utilities" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.204820 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerName="extract-utilities" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.205026 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b33aacfd-c55e-47a5-93cd-e426bba74235" containerName="registry-server" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.205704 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.208932 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd"] Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.211653 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.211859 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.250357 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ee73fd4-d24a-498b-b48a-aca42501618f-secret-volume\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.250472 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdvdg\" (UniqueName: \"kubernetes.io/projected/5ee73fd4-d24a-498b-b48a-aca42501618f-kube-api-access-gdvdg\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.250533 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ee73fd4-d24a-498b-b48a-aca42501618f-config-volume\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.351550 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdvdg\" (UniqueName: \"kubernetes.io/projected/5ee73fd4-d24a-498b-b48a-aca42501618f-kube-api-access-gdvdg\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.351630 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ee73fd4-d24a-498b-b48a-aca42501618f-config-volume\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.351689 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ee73fd4-d24a-498b-b48a-aca42501618f-secret-volume\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.352514 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ee73fd4-d24a-498b-b48a-aca42501618f-config-volume\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.358993 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ee73fd4-d24a-498b-b48a-aca42501618f-secret-volume\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.371849 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdvdg\" (UniqueName: \"kubernetes.io/projected/5ee73fd4-d24a-498b-b48a-aca42501618f-kube-api-access-gdvdg\") pod \"collect-profiles-29335185-7r4vd\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.524596 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:00 crc kubenswrapper[4788]: I1010 15:45:00.950969 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd"] Oct 10 15:45:01 crc kubenswrapper[4788]: I1010 15:45:01.234221 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:45:01 crc kubenswrapper[4788]: E1010 15:45:01.234456 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:45:01 crc kubenswrapper[4788]: I1010 15:45:01.808787 4788 generic.go:334] "Generic (PLEG): container finished" podID="5ee73fd4-d24a-498b-b48a-aca42501618f" containerID="b34b745cdb3e93eff8625aece80c287229820a4544266b653fcdd0b1cf04fd44" exitCode=0 Oct 10 15:45:01 crc kubenswrapper[4788]: I1010 15:45:01.808912 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" event={"ID":"5ee73fd4-d24a-498b-b48a-aca42501618f","Type":"ContainerDied","Data":"b34b745cdb3e93eff8625aece80c287229820a4544266b653fcdd0b1cf04fd44"} Oct 10 15:45:01 crc kubenswrapper[4788]: I1010 15:45:01.809113 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" event={"ID":"5ee73fd4-d24a-498b-b48a-aca42501618f","Type":"ContainerStarted","Data":"f2557480e8c22beb434dfc072740ca85b0dc0b952d3d9ec11c93b685b5d60f10"} Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.161326 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.197334 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ee73fd4-d24a-498b-b48a-aca42501618f-secret-volume\") pod \"5ee73fd4-d24a-498b-b48a-aca42501618f\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.197438 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdvdg\" (UniqueName: \"kubernetes.io/projected/5ee73fd4-d24a-498b-b48a-aca42501618f-kube-api-access-gdvdg\") pod \"5ee73fd4-d24a-498b-b48a-aca42501618f\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.197535 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ee73fd4-d24a-498b-b48a-aca42501618f-config-volume\") pod \"5ee73fd4-d24a-498b-b48a-aca42501618f\" (UID: \"5ee73fd4-d24a-498b-b48a-aca42501618f\") " Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.198219 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ee73fd4-d24a-498b-b48a-aca42501618f-config-volume" (OuterVolumeSpecName: "config-volume") pod "5ee73fd4-d24a-498b-b48a-aca42501618f" (UID: "5ee73fd4-d24a-498b-b48a-aca42501618f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.207396 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ee73fd4-d24a-498b-b48a-aca42501618f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5ee73fd4-d24a-498b-b48a-aca42501618f" (UID: "5ee73fd4-d24a-498b-b48a-aca42501618f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.207445 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ee73fd4-d24a-498b-b48a-aca42501618f-kube-api-access-gdvdg" (OuterVolumeSpecName: "kube-api-access-gdvdg") pod "5ee73fd4-d24a-498b-b48a-aca42501618f" (UID: "5ee73fd4-d24a-498b-b48a-aca42501618f"). InnerVolumeSpecName "kube-api-access-gdvdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.298549 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdvdg\" (UniqueName: \"kubernetes.io/projected/5ee73fd4-d24a-498b-b48a-aca42501618f-kube-api-access-gdvdg\") on node \"crc\" DevicePath \"\"" Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.298600 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5ee73fd4-d24a-498b-b48a-aca42501618f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.298617 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5ee73fd4-d24a-498b-b48a-aca42501618f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.825991 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" event={"ID":"5ee73fd4-d24a-498b-b48a-aca42501618f","Type":"ContainerDied","Data":"f2557480e8c22beb434dfc072740ca85b0dc0b952d3d9ec11c93b685b5d60f10"} Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.826163 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2557480e8c22beb434dfc072740ca85b0dc0b952d3d9ec11c93b685b5d60f10" Oct 10 15:45:03 crc kubenswrapper[4788]: I1010 15:45:03.826127 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd" Oct 10 15:45:04 crc kubenswrapper[4788]: I1010 15:45:04.232592 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn"] Oct 10 15:45:04 crc kubenswrapper[4788]: I1010 15:45:04.243053 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335140-89dcn"] Oct 10 15:45:06 crc kubenswrapper[4788]: I1010 15:45:06.248438 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a3c9bad-9fee-4dd0-b553-52cebeb42d12" path="/var/lib/kubelet/pods/9a3c9bad-9fee-4dd0-b553-52cebeb42d12/volumes" Oct 10 15:45:11 crc kubenswrapper[4788]: I1010 15:45:11.192250 4788 scope.go:117] "RemoveContainer" containerID="a1adbcca9500ed3d840078abe319997f98e311b25bf852506f5584458cc7ef55" Oct 10 15:45:15 crc kubenswrapper[4788]: I1010 15:45:15.234264 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:45:15 crc kubenswrapper[4788]: E1010 15:45:15.234772 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:45:29 crc kubenswrapper[4788]: I1010 15:45:29.234041 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:45:29 crc kubenswrapper[4788]: E1010 15:45:29.234990 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:45:40 crc kubenswrapper[4788]: I1010 15:45:40.233447 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:45:40 crc kubenswrapper[4788]: E1010 15:45:40.234244 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:45:52 crc kubenswrapper[4788]: I1010 15:45:52.234680 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:45:52 crc kubenswrapper[4788]: E1010 15:45:52.237025 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:46:03 crc kubenswrapper[4788]: I1010 15:46:03.233499 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:46:03 crc kubenswrapper[4788]: E1010 15:46:03.234265 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:46:14 crc kubenswrapper[4788]: I1010 15:46:14.233903 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:46:14 crc kubenswrapper[4788]: E1010 15:46:14.234775 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:46:27 crc kubenswrapper[4788]: I1010 15:46:27.234541 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:46:27 crc kubenswrapper[4788]: E1010 15:46:27.235462 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:46:41 crc kubenswrapper[4788]: I1010 15:46:41.234328 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:46:41 crc kubenswrapper[4788]: E1010 15:46:41.235643 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:46:55 crc kubenswrapper[4788]: I1010 15:46:55.235032 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:46:55 crc kubenswrapper[4788]: E1010 15:46:55.236370 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:47:07 crc kubenswrapper[4788]: I1010 15:47:07.233842 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:47:07 crc kubenswrapper[4788]: E1010 15:47:07.234337 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:47:19 crc kubenswrapper[4788]: I1010 15:47:19.233718 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:47:19 crc kubenswrapper[4788]: E1010 15:47:19.234659 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:47:33 crc kubenswrapper[4788]: I1010 15:47:33.233644 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:47:33 crc kubenswrapper[4788]: E1010 15:47:33.234236 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:47:47 crc kubenswrapper[4788]: I1010 15:47:47.234250 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:47:47 crc kubenswrapper[4788]: E1010 15:47:47.235242 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:47:58 crc kubenswrapper[4788]: I1010 15:47:58.240046 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:47:58 crc kubenswrapper[4788]: E1010 15:47:58.241100 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:48:11 crc kubenswrapper[4788]: I1010 15:48:11.233811 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:48:11 crc kubenswrapper[4788]: E1010 15:48:11.236035 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:48:25 crc kubenswrapper[4788]: I1010 15:48:25.235210 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:48:25 crc kubenswrapper[4788]: E1010 15:48:25.236632 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:48:36 crc kubenswrapper[4788]: I1010 15:48:36.234470 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:48:36 crc kubenswrapper[4788]: E1010 15:48:36.235492 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:48:48 crc kubenswrapper[4788]: I1010 15:48:48.242423 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:48:48 crc kubenswrapper[4788]: E1010 15:48:48.243344 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:48:59 crc kubenswrapper[4788]: I1010 15:48:59.234188 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:48:59 crc kubenswrapper[4788]: E1010 15:48:59.235070 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:49:11 crc kubenswrapper[4788]: I1010 15:49:11.233515 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:49:11 crc kubenswrapper[4788]: I1010 15:49:11.950255 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"ebb326f92f93c625a4898522de7d810733c488fd973ad91bd9295781b1c8e457"} Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.359076 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z2fnr"] Oct 10 15:51:09 crc kubenswrapper[4788]: E1010 15:51:09.360164 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ee73fd4-d24a-498b-b48a-aca42501618f" containerName="collect-profiles" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.360184 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ee73fd4-d24a-498b-b48a-aca42501618f" containerName="collect-profiles" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.360349 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ee73fd4-d24a-498b-b48a-aca42501618f" containerName="collect-profiles" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.361829 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.381483 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2fnr"] Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.455711 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-utilities\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.455830 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-catalog-content\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.455936 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl6jv\" (UniqueName: \"kubernetes.io/projected/73693dae-c506-46a2-99a7-1cd09b39b2d5-kube-api-access-zl6jv\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.558368 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-catalog-content\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.558577 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl6jv\" (UniqueName: \"kubernetes.io/projected/73693dae-c506-46a2-99a7-1cd09b39b2d5-kube-api-access-zl6jv\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.558693 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-utilities\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.559213 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-catalog-content\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.559668 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-utilities\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.610212 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl6jv\" (UniqueName: \"kubernetes.io/projected/73693dae-c506-46a2-99a7-1cd09b39b2d5-kube-api-access-zl6jv\") pod \"certified-operators-z2fnr\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:09 crc kubenswrapper[4788]: I1010 15:51:09.694096 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:10 crc kubenswrapper[4788]: I1010 15:51:10.001327 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2fnr"] Oct 10 15:51:10 crc kubenswrapper[4788]: I1010 15:51:10.066451 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2fnr" event={"ID":"73693dae-c506-46a2-99a7-1cd09b39b2d5","Type":"ContainerStarted","Data":"be412305fce2dc68f4c43f33c0fbf76b194f760785992cbf4013b332f388ccdd"} Oct 10 15:51:11 crc kubenswrapper[4788]: I1010 15:51:11.078296 4788 generic.go:334] "Generic (PLEG): container finished" podID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerID="3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d" exitCode=0 Oct 10 15:51:11 crc kubenswrapper[4788]: I1010 15:51:11.078379 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2fnr" event={"ID":"73693dae-c506-46a2-99a7-1cd09b39b2d5","Type":"ContainerDied","Data":"3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d"} Oct 10 15:51:11 crc kubenswrapper[4788]: I1010 15:51:11.081884 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 15:51:12 crc kubenswrapper[4788]: I1010 15:51:12.094303 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2fnr" event={"ID":"73693dae-c506-46a2-99a7-1cd09b39b2d5","Type":"ContainerStarted","Data":"09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7"} Oct 10 15:51:13 crc kubenswrapper[4788]: I1010 15:51:13.106845 4788 generic.go:334] "Generic (PLEG): container finished" podID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerID="09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7" exitCode=0 Oct 10 15:51:13 crc kubenswrapper[4788]: I1010 15:51:13.106926 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2fnr" event={"ID":"73693dae-c506-46a2-99a7-1cd09b39b2d5","Type":"ContainerDied","Data":"09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7"} Oct 10 15:51:14 crc kubenswrapper[4788]: I1010 15:51:14.118747 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2fnr" event={"ID":"73693dae-c506-46a2-99a7-1cd09b39b2d5","Type":"ContainerStarted","Data":"e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494"} Oct 10 15:51:19 crc kubenswrapper[4788]: I1010 15:51:19.695276 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:19 crc kubenswrapper[4788]: I1010 15:51:19.696581 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:19 crc kubenswrapper[4788]: I1010 15:51:19.780447 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:19 crc kubenswrapper[4788]: I1010 15:51:19.811357 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z2fnr" podStartSLOduration=8.390220871 podStartE2EDuration="10.811324095s" podCreationTimestamp="2025-10-10 15:51:09 +0000 UTC" firstStartedPulling="2025-10-10 15:51:11.081584575 +0000 UTC m=+3973.531300123" lastFinishedPulling="2025-10-10 15:51:13.502687759 +0000 UTC m=+3975.952403347" observedRunningTime="2025-10-10 15:51:14.140003865 +0000 UTC m=+3976.589719423" watchObservedRunningTime="2025-10-10 15:51:19.811324095 +0000 UTC m=+3982.261039673" Oct 10 15:51:20 crc kubenswrapper[4788]: I1010 15:51:20.246680 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:20 crc kubenswrapper[4788]: I1010 15:51:20.316251 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2fnr"] Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.200946 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z2fnr" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerName="registry-server" containerID="cri-o://e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494" gracePeriod=2 Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.624517 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.777675 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl6jv\" (UniqueName: \"kubernetes.io/projected/73693dae-c506-46a2-99a7-1cd09b39b2d5-kube-api-access-zl6jv\") pod \"73693dae-c506-46a2-99a7-1cd09b39b2d5\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.777728 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-catalog-content\") pod \"73693dae-c506-46a2-99a7-1cd09b39b2d5\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.777753 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-utilities\") pod \"73693dae-c506-46a2-99a7-1cd09b39b2d5\" (UID: \"73693dae-c506-46a2-99a7-1cd09b39b2d5\") " Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.780384 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-utilities" (OuterVolumeSpecName: "utilities") pod "73693dae-c506-46a2-99a7-1cd09b39b2d5" (UID: "73693dae-c506-46a2-99a7-1cd09b39b2d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.783079 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73693dae-c506-46a2-99a7-1cd09b39b2d5-kube-api-access-zl6jv" (OuterVolumeSpecName: "kube-api-access-zl6jv") pod "73693dae-c506-46a2-99a7-1cd09b39b2d5" (UID: "73693dae-c506-46a2-99a7-1cd09b39b2d5"). InnerVolumeSpecName "kube-api-access-zl6jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.825234 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73693dae-c506-46a2-99a7-1cd09b39b2d5" (UID: "73693dae-c506-46a2-99a7-1cd09b39b2d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.879806 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.879845 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl6jv\" (UniqueName: \"kubernetes.io/projected/73693dae-c506-46a2-99a7-1cd09b39b2d5-kube-api-access-zl6jv\") on node \"crc\" DevicePath \"\"" Oct 10 15:51:22 crc kubenswrapper[4788]: I1010 15:51:22.879857 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73693dae-c506-46a2-99a7-1cd09b39b2d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.218611 4788 generic.go:334] "Generic (PLEG): container finished" podID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerID="e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494" exitCode=0 Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.218696 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2fnr" event={"ID":"73693dae-c506-46a2-99a7-1cd09b39b2d5","Type":"ContainerDied","Data":"e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494"} Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.219310 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2fnr" event={"ID":"73693dae-c506-46a2-99a7-1cd09b39b2d5","Type":"ContainerDied","Data":"be412305fce2dc68f4c43f33c0fbf76b194f760785992cbf4013b332f388ccdd"} Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.219351 4788 scope.go:117] "RemoveContainer" containerID="e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.218823 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2fnr" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.251254 4788 scope.go:117] "RemoveContainer" containerID="09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.284216 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2fnr"] Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.294063 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z2fnr"] Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.302908 4788 scope.go:117] "RemoveContainer" containerID="3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.329950 4788 scope.go:117] "RemoveContainer" containerID="e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494" Oct 10 15:51:23 crc kubenswrapper[4788]: E1010 15:51:23.330470 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494\": container with ID starting with e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494 not found: ID does not exist" containerID="e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.330522 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494"} err="failed to get container status \"e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494\": rpc error: code = NotFound desc = could not find container \"e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494\": container with ID starting with e738d1cf408d0b3c51d603df40b81c86b7c4e31e394302b90707df89af50a494 not found: ID does not exist" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.330595 4788 scope.go:117] "RemoveContainer" containerID="09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7" Oct 10 15:51:23 crc kubenswrapper[4788]: E1010 15:51:23.330942 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7\": container with ID starting with 09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7 not found: ID does not exist" containerID="09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.331003 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7"} err="failed to get container status \"09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7\": rpc error: code = NotFound desc = could not find container \"09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7\": container with ID starting with 09f71678aa14934dd5db8f3c9630efef0e1ae1c383572702e077a9173b9501d7 not found: ID does not exist" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.331037 4788 scope.go:117] "RemoveContainer" containerID="3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d" Oct 10 15:51:23 crc kubenswrapper[4788]: E1010 15:51:23.331414 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d\": container with ID starting with 3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d not found: ID does not exist" containerID="3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d" Oct 10 15:51:23 crc kubenswrapper[4788]: I1010 15:51:23.331469 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d"} err="failed to get container status \"3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d\": rpc error: code = NotFound desc = could not find container \"3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d\": container with ID starting with 3fc421934cb7f2c0899bdb9dfcb6f9cfa15006f3dd091768ab47d621c609de0d not found: ID does not exist" Oct 10 15:51:24 crc kubenswrapper[4788]: I1010 15:51:24.250849 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" path="/var/lib/kubelet/pods/73693dae-c506-46a2-99a7-1cd09b39b2d5/volumes" Oct 10 15:51:29 crc kubenswrapper[4788]: I1010 15:51:29.407675 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:51:29 crc kubenswrapper[4788]: I1010 15:51:29.408312 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.565265 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sqfcd"] Oct 10 15:51:39 crc kubenswrapper[4788]: E1010 15:51:39.566392 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerName="extract-content" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.566405 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerName="extract-content" Oct 10 15:51:39 crc kubenswrapper[4788]: E1010 15:51:39.566418 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerName="registry-server" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.566426 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerName="registry-server" Oct 10 15:51:39 crc kubenswrapper[4788]: E1010 15:51:39.566451 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerName="extract-utilities" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.566458 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerName="extract-utilities" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.566592 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="73693dae-c506-46a2-99a7-1cd09b39b2d5" containerName="registry-server" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.567671 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.576183 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sqfcd"] Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.653550 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-utilities\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.653634 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llwhn\" (UniqueName: \"kubernetes.io/projected/0f73ccea-75e9-4c76-8701-917d127fcc76-kube-api-access-llwhn\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.653796 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-catalog-content\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.755544 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llwhn\" (UniqueName: \"kubernetes.io/projected/0f73ccea-75e9-4c76-8701-917d127fcc76-kube-api-access-llwhn\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.755619 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-catalog-content\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.755673 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-utilities\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.756204 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-utilities\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.756395 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-catalog-content\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.782030 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llwhn\" (UniqueName: \"kubernetes.io/projected/0f73ccea-75e9-4c76-8701-917d127fcc76-kube-api-access-llwhn\") pod \"redhat-operators-sqfcd\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:39 crc kubenswrapper[4788]: I1010 15:51:39.894900 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:40 crc kubenswrapper[4788]: I1010 15:51:40.159696 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sqfcd"] Oct 10 15:51:40 crc kubenswrapper[4788]: I1010 15:51:40.373603 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqfcd" event={"ID":"0f73ccea-75e9-4c76-8701-917d127fcc76","Type":"ContainerStarted","Data":"39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861"} Oct 10 15:51:40 crc kubenswrapper[4788]: I1010 15:51:40.373666 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqfcd" event={"ID":"0f73ccea-75e9-4c76-8701-917d127fcc76","Type":"ContainerStarted","Data":"00672849cc416c512b33018e43a5ef82d1df6e18c848e0de3478a965c64a0a92"} Oct 10 15:51:41 crc kubenswrapper[4788]: I1010 15:51:41.385601 4788 generic.go:334] "Generic (PLEG): container finished" podID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerID="39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861" exitCode=0 Oct 10 15:51:41 crc kubenswrapper[4788]: I1010 15:51:41.385672 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqfcd" event={"ID":"0f73ccea-75e9-4c76-8701-917d127fcc76","Type":"ContainerDied","Data":"39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861"} Oct 10 15:51:42 crc kubenswrapper[4788]: I1010 15:51:42.398880 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqfcd" event={"ID":"0f73ccea-75e9-4c76-8701-917d127fcc76","Type":"ContainerStarted","Data":"1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62"} Oct 10 15:51:43 crc kubenswrapper[4788]: I1010 15:51:43.409537 4788 generic.go:334] "Generic (PLEG): container finished" podID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerID="1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62" exitCode=0 Oct 10 15:51:43 crc kubenswrapper[4788]: I1010 15:51:43.409638 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqfcd" event={"ID":"0f73ccea-75e9-4c76-8701-917d127fcc76","Type":"ContainerDied","Data":"1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62"} Oct 10 15:51:44 crc kubenswrapper[4788]: I1010 15:51:44.422110 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqfcd" event={"ID":"0f73ccea-75e9-4c76-8701-917d127fcc76","Type":"ContainerStarted","Data":"7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a"} Oct 10 15:51:44 crc kubenswrapper[4788]: I1010 15:51:44.450975 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sqfcd" podStartSLOduration=3.007285243 podStartE2EDuration="5.450953569s" podCreationTimestamp="2025-10-10 15:51:39 +0000 UTC" firstStartedPulling="2025-10-10 15:51:41.390452696 +0000 UTC m=+4003.840168244" lastFinishedPulling="2025-10-10 15:51:43.834121022 +0000 UTC m=+4006.283836570" observedRunningTime="2025-10-10 15:51:44.446240332 +0000 UTC m=+4006.895955880" watchObservedRunningTime="2025-10-10 15:51:44.450953569 +0000 UTC m=+4006.900669117" Oct 10 15:51:49 crc kubenswrapper[4788]: I1010 15:51:49.897586 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:49 crc kubenswrapper[4788]: I1010 15:51:49.897858 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:49 crc kubenswrapper[4788]: I1010 15:51:49.954366 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:50 crc kubenswrapper[4788]: I1010 15:51:50.503576 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:50 crc kubenswrapper[4788]: I1010 15:51:50.559026 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sqfcd"] Oct 10 15:51:52 crc kubenswrapper[4788]: I1010 15:51:52.473780 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sqfcd" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerName="registry-server" containerID="cri-o://7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a" gracePeriod=2 Oct 10 15:51:52 crc kubenswrapper[4788]: I1010 15:51:52.825094 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:52 crc kubenswrapper[4788]: I1010 15:51:52.982828 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-catalog-content\") pod \"0f73ccea-75e9-4c76-8701-917d127fcc76\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " Oct 10 15:51:52 crc kubenswrapper[4788]: I1010 15:51:52.982878 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llwhn\" (UniqueName: \"kubernetes.io/projected/0f73ccea-75e9-4c76-8701-917d127fcc76-kube-api-access-llwhn\") pod \"0f73ccea-75e9-4c76-8701-917d127fcc76\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " Oct 10 15:51:52 crc kubenswrapper[4788]: I1010 15:51:52.982936 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-utilities\") pod \"0f73ccea-75e9-4c76-8701-917d127fcc76\" (UID: \"0f73ccea-75e9-4c76-8701-917d127fcc76\") " Oct 10 15:51:52 crc kubenswrapper[4788]: I1010 15:51:52.983949 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-utilities" (OuterVolumeSpecName: "utilities") pod "0f73ccea-75e9-4c76-8701-917d127fcc76" (UID: "0f73ccea-75e9-4c76-8701-917d127fcc76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:51:52 crc kubenswrapper[4788]: I1010 15:51:52.990342 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f73ccea-75e9-4c76-8701-917d127fcc76-kube-api-access-llwhn" (OuterVolumeSpecName: "kube-api-access-llwhn") pod "0f73ccea-75e9-4c76-8701-917d127fcc76" (UID: "0f73ccea-75e9-4c76-8701-917d127fcc76"). InnerVolumeSpecName "kube-api-access-llwhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.087571 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llwhn\" (UniqueName: \"kubernetes.io/projected/0f73ccea-75e9-4c76-8701-917d127fcc76-kube-api-access-llwhn\") on node \"crc\" DevicePath \"\"" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.089086 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.482723 4788 generic.go:334] "Generic (PLEG): container finished" podID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerID="7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a" exitCode=0 Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.482784 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqfcd" event={"ID":"0f73ccea-75e9-4c76-8701-917d127fcc76","Type":"ContainerDied","Data":"7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a"} Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.482845 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqfcd" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.483103 4788 scope.go:117] "RemoveContainer" containerID="7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.483089 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqfcd" event={"ID":"0f73ccea-75e9-4c76-8701-917d127fcc76","Type":"ContainerDied","Data":"00672849cc416c512b33018e43a5ef82d1df6e18c848e0de3478a965c64a0a92"} Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.499988 4788 scope.go:117] "RemoveContainer" containerID="1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.514726 4788 scope.go:117] "RemoveContainer" containerID="39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.552780 4788 scope.go:117] "RemoveContainer" containerID="7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a" Oct 10 15:51:53 crc kubenswrapper[4788]: E1010 15:51:53.553252 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a\": container with ID starting with 7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a not found: ID does not exist" containerID="7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.553309 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a"} err="failed to get container status \"7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a\": rpc error: code = NotFound desc = could not find container \"7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a\": container with ID starting with 7d8cb012684cb65d6eda34829e0f5cfce5969a2912095f91cf6494bdbabac98a not found: ID does not exist" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.553347 4788 scope.go:117] "RemoveContainer" containerID="1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62" Oct 10 15:51:53 crc kubenswrapper[4788]: E1010 15:51:53.553637 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62\": container with ID starting with 1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62 not found: ID does not exist" containerID="1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.553662 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62"} err="failed to get container status \"1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62\": rpc error: code = NotFound desc = could not find container \"1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62\": container with ID starting with 1e6787663c2e299d2d301714b957c35430bceedf353d5fbca2f3ae038500ad62 not found: ID does not exist" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.553676 4788 scope.go:117] "RemoveContainer" containerID="39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861" Oct 10 15:51:53 crc kubenswrapper[4788]: E1010 15:51:53.553945 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861\": container with ID starting with 39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861 not found: ID does not exist" containerID="39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.553968 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861"} err="failed to get container status \"39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861\": rpc error: code = NotFound desc = could not find container \"39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861\": container with ID starting with 39a80fedd126e2a78aba4f7b0f1f015468c51563e56feb73e25ce46491b73861 not found: ID does not exist" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.734676 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f73ccea-75e9-4c76-8701-917d127fcc76" (UID: "0f73ccea-75e9-4c76-8701-917d127fcc76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.798067 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f73ccea-75e9-4c76-8701-917d127fcc76-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.811846 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sqfcd"] Oct 10 15:51:53 crc kubenswrapper[4788]: I1010 15:51:53.821865 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sqfcd"] Oct 10 15:51:54 crc kubenswrapper[4788]: I1010 15:51:54.241778 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" path="/var/lib/kubelet/pods/0f73ccea-75e9-4c76-8701-917d127fcc76/volumes" Oct 10 15:51:59 crc kubenswrapper[4788]: I1010 15:51:59.406661 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:51:59 crc kubenswrapper[4788]: I1010 15:51:59.408238 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.407358 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.408297 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.408379 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.409669 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ebb326f92f93c625a4898522de7d810733c488fd973ad91bd9295781b1c8e457"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.409819 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://ebb326f92f93c625a4898522de7d810733c488fd973ad91bd9295781b1c8e457" gracePeriod=600 Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.810484 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="ebb326f92f93c625a4898522de7d810733c488fd973ad91bd9295781b1c8e457" exitCode=0 Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.810522 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"ebb326f92f93c625a4898522de7d810733c488fd973ad91bd9295781b1c8e457"} Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.810969 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7"} Oct 10 15:52:29 crc kubenswrapper[4788]: I1010 15:52:29.811003 4788 scope.go:117] "RemoveContainer" containerID="9f75a1648e3facfe611572e9c0fec4f41f0a6d38ec3e836949628ae2dcdbc9a5" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.210457 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jjz64"] Oct 10 15:54:17 crc kubenswrapper[4788]: E1010 15:54:17.212498 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerName="extract-content" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.212672 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerName="extract-content" Oct 10 15:54:17 crc kubenswrapper[4788]: E1010 15:54:17.212714 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerName="registry-server" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.212727 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerName="registry-server" Oct 10 15:54:17 crc kubenswrapper[4788]: E1010 15:54:17.212765 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerName="extract-utilities" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.212782 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerName="extract-utilities" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.213071 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f73ccea-75e9-4c76-8701-917d127fcc76" containerName="registry-server" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.220664 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.272503 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjz64"] Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.326907 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-catalog-content\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.327231 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwzbh\" (UniqueName: \"kubernetes.io/projected/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-kube-api-access-lwzbh\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.327379 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-utilities\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.429021 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-catalog-content\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.429102 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwzbh\" (UniqueName: \"kubernetes.io/projected/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-kube-api-access-lwzbh\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.429133 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-utilities\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.429667 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-utilities\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.429966 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-catalog-content\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.455973 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwzbh\" (UniqueName: \"kubernetes.io/projected/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-kube-api-access-lwzbh\") pod \"redhat-marketplace-jjz64\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.567355 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:17 crc kubenswrapper[4788]: I1010 15:54:17.791562 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjz64"] Oct 10 15:54:18 crc kubenswrapper[4788]: I1010 15:54:18.740112 4788 generic.go:334] "Generic (PLEG): container finished" podID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerID="2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9" exitCode=0 Oct 10 15:54:18 crc kubenswrapper[4788]: I1010 15:54:18.740176 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjz64" event={"ID":"93c125a9-368c-4ed2-92ea-d1854ad9ab4d","Type":"ContainerDied","Data":"2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9"} Oct 10 15:54:18 crc kubenswrapper[4788]: I1010 15:54:18.740204 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjz64" event={"ID":"93c125a9-368c-4ed2-92ea-d1854ad9ab4d","Type":"ContainerStarted","Data":"404a687fdb91b54359701a2ab18e84b20e8d736e92e4feacd549dea529f3b1eb"} Oct 10 15:54:19 crc kubenswrapper[4788]: I1010 15:54:19.750463 4788 generic.go:334] "Generic (PLEG): container finished" podID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerID="44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3" exitCode=0 Oct 10 15:54:19 crc kubenswrapper[4788]: I1010 15:54:19.750735 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjz64" event={"ID":"93c125a9-368c-4ed2-92ea-d1854ad9ab4d","Type":"ContainerDied","Data":"44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3"} Oct 10 15:54:20 crc kubenswrapper[4788]: I1010 15:54:20.767957 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjz64" event={"ID":"93c125a9-368c-4ed2-92ea-d1854ad9ab4d","Type":"ContainerStarted","Data":"7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff"} Oct 10 15:54:20 crc kubenswrapper[4788]: I1010 15:54:20.800204 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jjz64" podStartSLOduration=2.403755363 podStartE2EDuration="3.800175307s" podCreationTimestamp="2025-10-10 15:54:17 +0000 UTC" firstStartedPulling="2025-10-10 15:54:18.742397747 +0000 UTC m=+4161.192113305" lastFinishedPulling="2025-10-10 15:54:20.138817661 +0000 UTC m=+4162.588533249" observedRunningTime="2025-10-10 15:54:20.794096865 +0000 UTC m=+4163.243812413" watchObservedRunningTime="2025-10-10 15:54:20.800175307 +0000 UTC m=+4163.249890855" Oct 10 15:54:27 crc kubenswrapper[4788]: I1010 15:54:27.568109 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:27 crc kubenswrapper[4788]: I1010 15:54:27.568791 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:27 crc kubenswrapper[4788]: I1010 15:54:27.633933 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:27 crc kubenswrapper[4788]: I1010 15:54:27.903522 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:27 crc kubenswrapper[4788]: I1010 15:54:27.970388 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjz64"] Oct 10 15:54:29 crc kubenswrapper[4788]: I1010 15:54:29.406223 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:54:29 crc kubenswrapper[4788]: I1010 15:54:29.406295 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:54:29 crc kubenswrapper[4788]: I1010 15:54:29.854308 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jjz64" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerName="registry-server" containerID="cri-o://7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff" gracePeriod=2 Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.290852 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.379214 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwzbh\" (UniqueName: \"kubernetes.io/projected/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-kube-api-access-lwzbh\") pod \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.379460 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-catalog-content\") pod \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.379517 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-utilities\") pod \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\" (UID: \"93c125a9-368c-4ed2-92ea-d1854ad9ab4d\") " Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.381371 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-utilities" (OuterVolumeSpecName: "utilities") pod "93c125a9-368c-4ed2-92ea-d1854ad9ab4d" (UID: "93c125a9-368c-4ed2-92ea-d1854ad9ab4d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.397807 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93c125a9-368c-4ed2-92ea-d1854ad9ab4d" (UID: "93c125a9-368c-4ed2-92ea-d1854ad9ab4d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.400384 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-kube-api-access-lwzbh" (OuterVolumeSpecName: "kube-api-access-lwzbh") pod "93c125a9-368c-4ed2-92ea-d1854ad9ab4d" (UID: "93c125a9-368c-4ed2-92ea-d1854ad9ab4d"). InnerVolumeSpecName "kube-api-access-lwzbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.482497 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwzbh\" (UniqueName: \"kubernetes.io/projected/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-kube-api-access-lwzbh\") on node \"crc\" DevicePath \"\"" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.482579 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.482595 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93c125a9-368c-4ed2-92ea-d1854ad9ab4d-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.872559 4788 generic.go:334] "Generic (PLEG): container finished" podID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerID="7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff" exitCode=0 Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.872641 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjz64" event={"ID":"93c125a9-368c-4ed2-92ea-d1854ad9ab4d","Type":"ContainerDied","Data":"7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff"} Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.872737 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jjz64" event={"ID":"93c125a9-368c-4ed2-92ea-d1854ad9ab4d","Type":"ContainerDied","Data":"404a687fdb91b54359701a2ab18e84b20e8d736e92e4feacd549dea529f3b1eb"} Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.872776 4788 scope.go:117] "RemoveContainer" containerID="7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.872783 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jjz64" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.929409 4788 scope.go:117] "RemoveContainer" containerID="44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.955760 4788 scope.go:117] "RemoveContainer" containerID="2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.956850 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjz64"] Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.970384 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jjz64"] Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.983792 4788 scope.go:117] "RemoveContainer" containerID="7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff" Oct 10 15:54:30 crc kubenswrapper[4788]: E1010 15:54:30.984736 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff\": container with ID starting with 7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff not found: ID does not exist" containerID="7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.984793 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff"} err="failed to get container status \"7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff\": rpc error: code = NotFound desc = could not find container \"7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff\": container with ID starting with 7380325faf6f7f1583171ae976c673294ff08640860948fe89a94501108076ff not found: ID does not exist" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.984822 4788 scope.go:117] "RemoveContainer" containerID="44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3" Oct 10 15:54:30 crc kubenswrapper[4788]: E1010 15:54:30.985465 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3\": container with ID starting with 44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3 not found: ID does not exist" containerID="44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.985503 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3"} err="failed to get container status \"44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3\": rpc error: code = NotFound desc = could not find container \"44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3\": container with ID starting with 44ce810f2d5ce23aa03f329a8d4df49f9b847941e282939b85f5a8c8fd6807e3 not found: ID does not exist" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.985530 4788 scope.go:117] "RemoveContainer" containerID="2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9" Oct 10 15:54:30 crc kubenswrapper[4788]: E1010 15:54:30.985993 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9\": container with ID starting with 2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9 not found: ID does not exist" containerID="2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9" Oct 10 15:54:30 crc kubenswrapper[4788]: I1010 15:54:30.986092 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9"} err="failed to get container status \"2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9\": rpc error: code = NotFound desc = could not find container \"2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9\": container with ID starting with 2638f60da5b392b50195b3c72f86cf8f99c18941e70dd9118ecdfb5fd32e7fe9 not found: ID does not exist" Oct 10 15:54:32 crc kubenswrapper[4788]: I1010 15:54:32.249630 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" path="/var/lib/kubelet/pods/93c125a9-368c-4ed2-92ea-d1854ad9ab4d/volumes" Oct 10 15:54:59 crc kubenswrapper[4788]: I1010 15:54:59.405988 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:54:59 crc kubenswrapper[4788]: I1010 15:54:59.408121 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.610260 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4rzkv"] Oct 10 15:55:17 crc kubenswrapper[4788]: E1010 15:55:17.611638 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerName="extract-content" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.611671 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerName="extract-content" Oct 10 15:55:17 crc kubenswrapper[4788]: E1010 15:55:17.611760 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerName="extract-utilities" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.611778 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerName="extract-utilities" Oct 10 15:55:17 crc kubenswrapper[4788]: E1010 15:55:17.611813 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerName="registry-server" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.611833 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerName="registry-server" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.612218 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c125a9-368c-4ed2-92ea-d1854ad9ab4d" containerName="registry-server" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.618659 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.623602 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4rzkv"] Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.690577 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-utilities\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.690655 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98v7r\" (UniqueName: \"kubernetes.io/projected/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-kube-api-access-98v7r\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.690731 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-catalog-content\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.792085 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-utilities\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.792174 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98v7r\" (UniqueName: \"kubernetes.io/projected/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-kube-api-access-98v7r\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.792242 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-catalog-content\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.792901 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-catalog-content\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.792925 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-utilities\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.813222 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98v7r\" (UniqueName: \"kubernetes.io/projected/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-kube-api-access-98v7r\") pod \"community-operators-4rzkv\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:17 crc kubenswrapper[4788]: I1010 15:55:17.941837 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:18 crc kubenswrapper[4788]: I1010 15:55:18.365572 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4rzkv"] Oct 10 15:55:19 crc kubenswrapper[4788]: I1010 15:55:19.322285 4788 generic.go:334] "Generic (PLEG): container finished" podID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerID="f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5" exitCode=0 Oct 10 15:55:19 crc kubenswrapper[4788]: I1010 15:55:19.322374 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rzkv" event={"ID":"a9b97bb5-9242-4d5e-b5df-e2fcb087259f","Type":"ContainerDied","Data":"f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5"} Oct 10 15:55:19 crc kubenswrapper[4788]: I1010 15:55:19.322706 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rzkv" event={"ID":"a9b97bb5-9242-4d5e-b5df-e2fcb087259f","Type":"ContainerStarted","Data":"3f89988f53c28850ce7b9595cd1b953b020b096acbfdeb28742af5c833113191"} Oct 10 15:55:21 crc kubenswrapper[4788]: I1010 15:55:21.342987 4788 generic.go:334] "Generic (PLEG): container finished" podID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerID="b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567" exitCode=0 Oct 10 15:55:21 crc kubenswrapper[4788]: I1010 15:55:21.345411 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rzkv" event={"ID":"a9b97bb5-9242-4d5e-b5df-e2fcb087259f","Type":"ContainerDied","Data":"b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567"} Oct 10 15:55:22 crc kubenswrapper[4788]: I1010 15:55:22.353411 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rzkv" event={"ID":"a9b97bb5-9242-4d5e-b5df-e2fcb087259f","Type":"ContainerStarted","Data":"3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd"} Oct 10 15:55:22 crc kubenswrapper[4788]: I1010 15:55:22.378194 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4rzkv" podStartSLOduration=2.953358196 podStartE2EDuration="5.378175772s" podCreationTimestamp="2025-10-10 15:55:17 +0000 UTC" firstStartedPulling="2025-10-10 15:55:19.325859674 +0000 UTC m=+4221.775575252" lastFinishedPulling="2025-10-10 15:55:21.75067728 +0000 UTC m=+4224.200392828" observedRunningTime="2025-10-10 15:55:22.373228159 +0000 UTC m=+4224.822943707" watchObservedRunningTime="2025-10-10 15:55:22.378175772 +0000 UTC m=+4224.827891320" Oct 10 15:55:27 crc kubenswrapper[4788]: I1010 15:55:27.942465 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:27 crc kubenswrapper[4788]: I1010 15:55:27.943013 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:27 crc kubenswrapper[4788]: I1010 15:55:27.982216 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:28 crc kubenswrapper[4788]: I1010 15:55:28.459419 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:28 crc kubenswrapper[4788]: I1010 15:55:28.505928 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4rzkv"] Oct 10 15:55:29 crc kubenswrapper[4788]: I1010 15:55:29.406741 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 15:55:29 crc kubenswrapper[4788]: I1010 15:55:29.406814 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 15:55:29 crc kubenswrapper[4788]: I1010 15:55:29.406871 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 15:55:29 crc kubenswrapper[4788]: I1010 15:55:29.407719 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 15:55:29 crc kubenswrapper[4788]: I1010 15:55:29.407828 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" gracePeriod=600 Oct 10 15:55:29 crc kubenswrapper[4788]: E1010 15:55:29.559316 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.437399 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" exitCode=0 Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.437506 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7"} Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.437827 4788 scope.go:117] "RemoveContainer" containerID="ebb326f92f93c625a4898522de7d810733c488fd973ad91bd9295781b1c8e457" Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.437957 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4rzkv" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerName="registry-server" containerID="cri-o://3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd" gracePeriod=2 Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.438987 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:55:30 crc kubenswrapper[4788]: E1010 15:55:30.439580 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.857231 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.996479 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-utilities\") pod \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.997330 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98v7r\" (UniqueName: \"kubernetes.io/projected/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-kube-api-access-98v7r\") pod \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.997385 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-catalog-content\") pod \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\" (UID: \"a9b97bb5-9242-4d5e-b5df-e2fcb087259f\") " Oct 10 15:55:30 crc kubenswrapper[4788]: I1010 15:55:30.998739 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-utilities" (OuterVolumeSpecName: "utilities") pod "a9b97bb5-9242-4d5e-b5df-e2fcb087259f" (UID: "a9b97bb5-9242-4d5e-b5df-e2fcb087259f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.005664 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-kube-api-access-98v7r" (OuterVolumeSpecName: "kube-api-access-98v7r") pod "a9b97bb5-9242-4d5e-b5df-e2fcb087259f" (UID: "a9b97bb5-9242-4d5e-b5df-e2fcb087259f"). InnerVolumeSpecName "kube-api-access-98v7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.050422 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9b97bb5-9242-4d5e-b5df-e2fcb087259f" (UID: "a9b97bb5-9242-4d5e-b5df-e2fcb087259f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.099061 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.099100 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98v7r\" (UniqueName: \"kubernetes.io/projected/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-kube-api-access-98v7r\") on node \"crc\" DevicePath \"\"" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.099115 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9b97bb5-9242-4d5e-b5df-e2fcb087259f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.447868 4788 generic.go:334] "Generic (PLEG): container finished" podID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerID="3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd" exitCode=0 Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.447986 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rzkv" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.450330 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rzkv" event={"ID":"a9b97bb5-9242-4d5e-b5df-e2fcb087259f","Type":"ContainerDied","Data":"3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd"} Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.450623 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rzkv" event={"ID":"a9b97bb5-9242-4d5e-b5df-e2fcb087259f","Type":"ContainerDied","Data":"3f89988f53c28850ce7b9595cd1b953b020b096acbfdeb28742af5c833113191"} Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.450705 4788 scope.go:117] "RemoveContainer" containerID="3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.474526 4788 scope.go:117] "RemoveContainer" containerID="b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.497270 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4rzkv"] Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.506118 4788 scope.go:117] "RemoveContainer" containerID="f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.510499 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4rzkv"] Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.536215 4788 scope.go:117] "RemoveContainer" containerID="3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd" Oct 10 15:55:31 crc kubenswrapper[4788]: E1010 15:55:31.536627 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd\": container with ID starting with 3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd not found: ID does not exist" containerID="3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.536680 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd"} err="failed to get container status \"3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd\": rpc error: code = NotFound desc = could not find container \"3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd\": container with ID starting with 3d4547d317b34519bfeb2e2f63ba94da955d93c31eb268a1885ab313d62cbbfd not found: ID does not exist" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.536718 4788 scope.go:117] "RemoveContainer" containerID="b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567" Oct 10 15:55:31 crc kubenswrapper[4788]: E1010 15:55:31.536946 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567\": container with ID starting with b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567 not found: ID does not exist" containerID="b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.536969 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567"} err="failed to get container status \"b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567\": rpc error: code = NotFound desc = could not find container \"b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567\": container with ID starting with b0d0ca55256b7cd54c24ab323ded8499d587539ba57758b0a8beafab0dcf8567 not found: ID does not exist" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.536983 4788 scope.go:117] "RemoveContainer" containerID="f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5" Oct 10 15:55:31 crc kubenswrapper[4788]: E1010 15:55:31.537405 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5\": container with ID starting with f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5 not found: ID does not exist" containerID="f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5" Oct 10 15:55:31 crc kubenswrapper[4788]: I1010 15:55:31.537438 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5"} err="failed to get container status \"f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5\": rpc error: code = NotFound desc = could not find container \"f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5\": container with ID starting with f5d826d2aafba803bd878c4e6ee3cebca08e4db4d7d0ff8bde294a0725a29ac5 not found: ID does not exist" Oct 10 15:55:32 crc kubenswrapper[4788]: I1010 15:55:32.248017 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" path="/var/lib/kubelet/pods/a9b97bb5-9242-4d5e-b5df-e2fcb087259f/volumes" Oct 10 15:55:42 crc kubenswrapper[4788]: I1010 15:55:42.233951 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:55:42 crc kubenswrapper[4788]: E1010 15:55:42.235029 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:55:55 crc kubenswrapper[4788]: I1010 15:55:55.233505 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:55:55 crc kubenswrapper[4788]: E1010 15:55:55.234195 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:56:08 crc kubenswrapper[4788]: I1010 15:56:08.245663 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:56:08 crc kubenswrapper[4788]: E1010 15:56:08.246911 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:56:19 crc kubenswrapper[4788]: I1010 15:56:19.235126 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:56:19 crc kubenswrapper[4788]: E1010 15:56:19.236263 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:56:33 crc kubenswrapper[4788]: I1010 15:56:33.234126 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:56:33 crc kubenswrapper[4788]: E1010 15:56:33.235244 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:56:46 crc kubenswrapper[4788]: I1010 15:56:46.234347 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:56:46 crc kubenswrapper[4788]: E1010 15:56:46.234975 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:56:57 crc kubenswrapper[4788]: I1010 15:56:57.234322 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:56:57 crc kubenswrapper[4788]: E1010 15:56:57.234827 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:57:09 crc kubenswrapper[4788]: I1010 15:57:09.234803 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:57:09 crc kubenswrapper[4788]: E1010 15:57:09.235877 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:57:22 crc kubenswrapper[4788]: I1010 15:57:22.235444 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:57:22 crc kubenswrapper[4788]: E1010 15:57:22.236570 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:57:33 crc kubenswrapper[4788]: I1010 15:57:33.233846 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:57:33 crc kubenswrapper[4788]: E1010 15:57:33.234610 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:57:44 crc kubenswrapper[4788]: I1010 15:57:44.234341 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:57:44 crc kubenswrapper[4788]: E1010 15:57:44.235519 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:57:57 crc kubenswrapper[4788]: I1010 15:57:57.232996 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:57:57 crc kubenswrapper[4788]: E1010 15:57:57.233577 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:58:11 crc kubenswrapper[4788]: I1010 15:58:11.234291 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:58:11 crc kubenswrapper[4788]: E1010 15:58:11.236060 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:58:24 crc kubenswrapper[4788]: I1010 15:58:24.234197 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:58:24 crc kubenswrapper[4788]: E1010 15:58:24.234949 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:58:35 crc kubenswrapper[4788]: I1010 15:58:35.233598 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:58:35 crc kubenswrapper[4788]: E1010 15:58:35.234411 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:58:46 crc kubenswrapper[4788]: I1010 15:58:46.234001 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:58:46 crc kubenswrapper[4788]: E1010 15:58:46.234864 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:58:57 crc kubenswrapper[4788]: I1010 15:58:57.234097 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:58:57 crc kubenswrapper[4788]: E1010 15:58:57.235129 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:59:12 crc kubenswrapper[4788]: I1010 15:59:12.234793 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:59:12 crc kubenswrapper[4788]: E1010 15:59:12.236073 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:59:24 crc kubenswrapper[4788]: I1010 15:59:24.234600 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:59:24 crc kubenswrapper[4788]: E1010 15:59:24.235192 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:59:35 crc kubenswrapper[4788]: I1010 15:59:35.234570 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:59:35 crc kubenswrapper[4788]: E1010 15:59:35.237171 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 15:59:48 crc kubenswrapper[4788]: I1010 15:59:48.241618 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 15:59:48 crc kubenswrapper[4788]: E1010 15:59:48.242870 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.158905 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw"] Oct 10 16:00:00 crc kubenswrapper[4788]: E1010 16:00:00.159784 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerName="registry-server" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.159803 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerName="registry-server" Oct 10 16:00:00 crc kubenswrapper[4788]: E1010 16:00:00.159831 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerName="extract-utilities" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.159838 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerName="extract-utilities" Oct 10 16:00:00 crc kubenswrapper[4788]: E1010 16:00:00.159858 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerName="extract-content" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.159865 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerName="extract-content" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.160087 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9b97bb5-9242-4d5e-b5df-e2fcb087259f" containerName="registry-server" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.160817 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.163818 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.168307 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.173451 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw"] Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.287925 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlqql\" (UniqueName: \"kubernetes.io/projected/035cfcc7-90ca-4547-b3b7-2094e8ff225e-kube-api-access-zlqql\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.287979 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/035cfcc7-90ca-4547-b3b7-2094e8ff225e-config-volume\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.288050 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/035cfcc7-90ca-4547-b3b7-2094e8ff225e-secret-volume\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.389435 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/035cfcc7-90ca-4547-b3b7-2094e8ff225e-config-volume\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.389587 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/035cfcc7-90ca-4547-b3b7-2094e8ff225e-secret-volume\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.389675 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlqql\" (UniqueName: \"kubernetes.io/projected/035cfcc7-90ca-4547-b3b7-2094e8ff225e-kube-api-access-zlqql\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.392332 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/035cfcc7-90ca-4547-b3b7-2094e8ff225e-config-volume\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.402188 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/035cfcc7-90ca-4547-b3b7-2094e8ff225e-secret-volume\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.413098 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlqql\" (UniqueName: \"kubernetes.io/projected/035cfcc7-90ca-4547-b3b7-2094e8ff225e-kube-api-access-zlqql\") pod \"collect-profiles-29335200-pbnlw\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:00 crc kubenswrapper[4788]: I1010 16:00:00.490820 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:01 crc kubenswrapper[4788]: I1010 16:00:01.003516 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw"] Oct 10 16:00:01 crc kubenswrapper[4788]: I1010 16:00:01.807385 4788 generic.go:334] "Generic (PLEG): container finished" podID="035cfcc7-90ca-4547-b3b7-2094e8ff225e" containerID="9108b8823a0aeda1eeb999175ae7707a71c827e6afc666b06281dd814e69d626" exitCode=0 Oct 10 16:00:01 crc kubenswrapper[4788]: I1010 16:00:01.807867 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" event={"ID":"035cfcc7-90ca-4547-b3b7-2094e8ff225e","Type":"ContainerDied","Data":"9108b8823a0aeda1eeb999175ae7707a71c827e6afc666b06281dd814e69d626"} Oct 10 16:00:01 crc kubenswrapper[4788]: I1010 16:00:01.807916 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" event={"ID":"035cfcc7-90ca-4547-b3b7-2094e8ff225e","Type":"ContainerStarted","Data":"108a6b71949d870c7bafd8225f2bc12d944496b1d924d5162e0dbfedbb652f24"} Oct 10 16:00:02 crc kubenswrapper[4788]: I1010 16:00:02.234465 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 16:00:02 crc kubenswrapper[4788]: E1010 16:00:02.234699 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.260806 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.341587 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/035cfcc7-90ca-4547-b3b7-2094e8ff225e-secret-volume\") pod \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.341647 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/035cfcc7-90ca-4547-b3b7-2094e8ff225e-config-volume\") pod \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.341791 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlqql\" (UniqueName: \"kubernetes.io/projected/035cfcc7-90ca-4547-b3b7-2094e8ff225e-kube-api-access-zlqql\") pod \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\" (UID: \"035cfcc7-90ca-4547-b3b7-2094e8ff225e\") " Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.343330 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/035cfcc7-90ca-4547-b3b7-2094e8ff225e-config-volume" (OuterVolumeSpecName: "config-volume") pod "035cfcc7-90ca-4547-b3b7-2094e8ff225e" (UID: "035cfcc7-90ca-4547-b3b7-2094e8ff225e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.351386 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/035cfcc7-90ca-4547-b3b7-2094e8ff225e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "035cfcc7-90ca-4547-b3b7-2094e8ff225e" (UID: "035cfcc7-90ca-4547-b3b7-2094e8ff225e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.351657 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/035cfcc7-90ca-4547-b3b7-2094e8ff225e-kube-api-access-zlqql" (OuterVolumeSpecName: "kube-api-access-zlqql") pod "035cfcc7-90ca-4547-b3b7-2094e8ff225e" (UID: "035cfcc7-90ca-4547-b3b7-2094e8ff225e"). InnerVolumeSpecName "kube-api-access-zlqql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.444574 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/035cfcc7-90ca-4547-b3b7-2094e8ff225e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.444626 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/035cfcc7-90ca-4547-b3b7-2094e8ff225e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.444649 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlqql\" (UniqueName: \"kubernetes.io/projected/035cfcc7-90ca-4547-b3b7-2094e8ff225e-kube-api-access-zlqql\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.830438 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" event={"ID":"035cfcc7-90ca-4547-b3b7-2094e8ff225e","Type":"ContainerDied","Data":"108a6b71949d870c7bafd8225f2bc12d944496b1d924d5162e0dbfedbb652f24"} Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.830494 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="108a6b71949d870c7bafd8225f2bc12d944496b1d924d5162e0dbfedbb652f24" Oct 10 16:00:03 crc kubenswrapper[4788]: I1010 16:00:03.831058 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw" Oct 10 16:00:04 crc kubenswrapper[4788]: I1010 16:00:04.360054 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s"] Oct 10 16:00:04 crc kubenswrapper[4788]: I1010 16:00:04.366668 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335155-kqc4s"] Oct 10 16:00:06 crc kubenswrapper[4788]: I1010 16:00:06.249496 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f5943a2-fc93-4c06-bd9e-bddfafa70008" path="/var/lib/kubelet/pods/1f5943a2-fc93-4c06-bd9e-bddfafa70008/volumes" Oct 10 16:00:11 crc kubenswrapper[4788]: I1010 16:00:11.602573 4788 scope.go:117] "RemoveContainer" containerID="ab1d3fe6a8c266881d42fd2b1a9fbe4f5112d4b810842047600cbe7d2a71dffa" Oct 10 16:00:14 crc kubenswrapper[4788]: I1010 16:00:14.233986 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 16:00:14 crc kubenswrapper[4788]: E1010 16:00:14.234541 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.574378 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-8442z"] Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.583014 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-8442z"] Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.752972 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-krsh9"] Oct 10 16:00:18 crc kubenswrapper[4788]: E1010 16:00:18.753456 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="035cfcc7-90ca-4547-b3b7-2094e8ff225e" containerName="collect-profiles" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.753475 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="035cfcc7-90ca-4547-b3b7-2094e8ff225e" containerName="collect-profiles" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.753689 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="035cfcc7-90ca-4547-b3b7-2094e8ff225e" containerName="collect-profiles" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.754424 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.757489 4788 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-psrrw" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.757802 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.757975 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.760483 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.766859 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-krsh9"] Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.913328 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2jdf\" (UniqueName: \"kubernetes.io/projected/dd7555cf-8c7d-4273-83c1-3195779fff07-kube-api-access-b2jdf\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.913988 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dd7555cf-8c7d-4273-83c1-3195779fff07-node-mnt\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:18 crc kubenswrapper[4788]: I1010 16:00:18.914428 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dd7555cf-8c7d-4273-83c1-3195779fff07-crc-storage\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.015724 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dd7555cf-8c7d-4273-83c1-3195779fff07-node-mnt\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.015784 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dd7555cf-8c7d-4273-83c1-3195779fff07-crc-storage\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.015866 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2jdf\" (UniqueName: \"kubernetes.io/projected/dd7555cf-8c7d-4273-83c1-3195779fff07-kube-api-access-b2jdf\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.016112 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dd7555cf-8c7d-4273-83c1-3195779fff07-node-mnt\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.017497 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dd7555cf-8c7d-4273-83c1-3195779fff07-crc-storage\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.038619 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2jdf\" (UniqueName: \"kubernetes.io/projected/dd7555cf-8c7d-4273-83c1-3195779fff07-kube-api-access-b2jdf\") pod \"crc-storage-crc-krsh9\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.084342 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.372206 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-krsh9"] Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.380439 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:00:19 crc kubenswrapper[4788]: I1010 16:00:19.988832 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-krsh9" event={"ID":"dd7555cf-8c7d-4273-83c1-3195779fff07","Type":"ContainerStarted","Data":"2c97bc2858047ef428a12d8f181e34c2c729bebedf2dc5e9d0475ba3fe28d77b"} Oct 10 16:00:20 crc kubenswrapper[4788]: I1010 16:00:20.242388 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d08229c-455b-4b6a-a904-f114306b9e1b" path="/var/lib/kubelet/pods/1d08229c-455b-4b6a-a904-f114306b9e1b/volumes" Oct 10 16:00:20 crc kubenswrapper[4788]: I1010 16:00:20.996398 4788 generic.go:334] "Generic (PLEG): container finished" podID="dd7555cf-8c7d-4273-83c1-3195779fff07" containerID="edc563279ab190b7eabed0297e85d99da4dfa732ddbbcf2c0cd277a319ef7ba3" exitCode=0 Oct 10 16:00:20 crc kubenswrapper[4788]: I1010 16:00:20.996478 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-krsh9" event={"ID":"dd7555cf-8c7d-4273-83c1-3195779fff07","Type":"ContainerDied","Data":"edc563279ab190b7eabed0297e85d99da4dfa732ddbbcf2c0cd277a319ef7ba3"} Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.284116 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.470243 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dd7555cf-8c7d-4273-83c1-3195779fff07-crc-storage\") pod \"dd7555cf-8c7d-4273-83c1-3195779fff07\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.470337 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dd7555cf-8c7d-4273-83c1-3195779fff07-node-mnt\") pod \"dd7555cf-8c7d-4273-83c1-3195779fff07\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.470471 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2jdf\" (UniqueName: \"kubernetes.io/projected/dd7555cf-8c7d-4273-83c1-3195779fff07-kube-api-access-b2jdf\") pod \"dd7555cf-8c7d-4273-83c1-3195779fff07\" (UID: \"dd7555cf-8c7d-4273-83c1-3195779fff07\") " Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.470529 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dd7555cf-8c7d-4273-83c1-3195779fff07-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "dd7555cf-8c7d-4273-83c1-3195779fff07" (UID: "dd7555cf-8c7d-4273-83c1-3195779fff07"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.470974 4788 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dd7555cf-8c7d-4273-83c1-3195779fff07-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.476296 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd7555cf-8c7d-4273-83c1-3195779fff07-kube-api-access-b2jdf" (OuterVolumeSpecName: "kube-api-access-b2jdf") pod "dd7555cf-8c7d-4273-83c1-3195779fff07" (UID: "dd7555cf-8c7d-4273-83c1-3195779fff07"). InnerVolumeSpecName "kube-api-access-b2jdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.489467 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd7555cf-8c7d-4273-83c1-3195779fff07-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "dd7555cf-8c7d-4273-83c1-3195779fff07" (UID: "dd7555cf-8c7d-4273-83c1-3195779fff07"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.572469 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2jdf\" (UniqueName: \"kubernetes.io/projected/dd7555cf-8c7d-4273-83c1-3195779fff07-kube-api-access-b2jdf\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:22 crc kubenswrapper[4788]: I1010 16:00:22.572526 4788 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dd7555cf-8c7d-4273-83c1-3195779fff07-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:23 crc kubenswrapper[4788]: I1010 16:00:23.015896 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-krsh9" event={"ID":"dd7555cf-8c7d-4273-83c1-3195779fff07","Type":"ContainerDied","Data":"2c97bc2858047ef428a12d8f181e34c2c729bebedf2dc5e9d0475ba3fe28d77b"} Oct 10 16:00:23 crc kubenswrapper[4788]: I1010 16:00:23.015968 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c97bc2858047ef428a12d8f181e34c2c729bebedf2dc5e9d0475ba3fe28d77b" Oct 10 16:00:23 crc kubenswrapper[4788]: I1010 16:00:23.015962 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-krsh9" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.782757 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-krsh9"] Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.788940 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-krsh9"] Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.931918 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-4zd6c"] Oct 10 16:00:24 crc kubenswrapper[4788]: E1010 16:00:24.932352 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7555cf-8c7d-4273-83c1-3195779fff07" containerName="storage" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.932373 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7555cf-8c7d-4273-83c1-3195779fff07" containerName="storage" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.932561 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd7555cf-8c7d-4273-83c1-3195779fff07" containerName="storage" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.933235 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.935337 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.936972 4788 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-psrrw" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.937019 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.938169 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 10 16:00:24 crc kubenswrapper[4788]: I1010 16:00:24.940046 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4zd6c"] Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.116129 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1181499d-f129-4c48-8376-f01992cc12af-node-mnt\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.116395 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1181499d-f129-4c48-8376-f01992cc12af-crc-storage\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.116519 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2vtx\" (UniqueName: \"kubernetes.io/projected/1181499d-f129-4c48-8376-f01992cc12af-kube-api-access-h2vtx\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.218327 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1181499d-f129-4c48-8376-f01992cc12af-crc-storage\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.218424 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2vtx\" (UniqueName: \"kubernetes.io/projected/1181499d-f129-4c48-8376-f01992cc12af-kube-api-access-h2vtx\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.218558 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1181499d-f129-4c48-8376-f01992cc12af-node-mnt\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.218920 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1181499d-f129-4c48-8376-f01992cc12af-node-mnt\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.219302 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1181499d-f129-4c48-8376-f01992cc12af-crc-storage\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.234442 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 16:00:25 crc kubenswrapper[4788]: E1010 16:00:25.234735 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.236621 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2vtx\" (UniqueName: \"kubernetes.io/projected/1181499d-f129-4c48-8376-f01992cc12af-kube-api-access-h2vtx\") pod \"crc-storage-crc-4zd6c\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.258471 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:25 crc kubenswrapper[4788]: I1010 16:00:25.751664 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-4zd6c"] Oct 10 16:00:26 crc kubenswrapper[4788]: I1010 16:00:26.044534 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4zd6c" event={"ID":"1181499d-f129-4c48-8376-f01992cc12af","Type":"ContainerStarted","Data":"e949139ebbdc6105b26b7139d59090c145971b1f5766457a2fcc982e6915a19b"} Oct 10 16:00:26 crc kubenswrapper[4788]: I1010 16:00:26.245216 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd7555cf-8c7d-4273-83c1-3195779fff07" path="/var/lib/kubelet/pods/dd7555cf-8c7d-4273-83c1-3195779fff07/volumes" Oct 10 16:00:27 crc kubenswrapper[4788]: I1010 16:00:27.054295 4788 generic.go:334] "Generic (PLEG): container finished" podID="1181499d-f129-4c48-8376-f01992cc12af" containerID="d625f7407bbc9d987aa29d5b13211c41ec575cd73867e86eba640a574e0524fe" exitCode=0 Oct 10 16:00:27 crc kubenswrapper[4788]: I1010 16:00:27.054409 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4zd6c" event={"ID":"1181499d-f129-4c48-8376-f01992cc12af","Type":"ContainerDied","Data":"d625f7407bbc9d987aa29d5b13211c41ec575cd73867e86eba640a574e0524fe"} Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.408068 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.567846 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1181499d-f129-4c48-8376-f01992cc12af-node-mnt\") pod \"1181499d-f129-4c48-8376-f01992cc12af\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.567955 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1181499d-f129-4c48-8376-f01992cc12af-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "1181499d-f129-4c48-8376-f01992cc12af" (UID: "1181499d-f129-4c48-8376-f01992cc12af"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.567978 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2vtx\" (UniqueName: \"kubernetes.io/projected/1181499d-f129-4c48-8376-f01992cc12af-kube-api-access-h2vtx\") pod \"1181499d-f129-4c48-8376-f01992cc12af\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.568132 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1181499d-f129-4c48-8376-f01992cc12af-crc-storage\") pod \"1181499d-f129-4c48-8376-f01992cc12af\" (UID: \"1181499d-f129-4c48-8376-f01992cc12af\") " Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.568841 4788 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1181499d-f129-4c48-8376-f01992cc12af-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.575440 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1181499d-f129-4c48-8376-f01992cc12af-kube-api-access-h2vtx" (OuterVolumeSpecName: "kube-api-access-h2vtx") pod "1181499d-f129-4c48-8376-f01992cc12af" (UID: "1181499d-f129-4c48-8376-f01992cc12af"). InnerVolumeSpecName "kube-api-access-h2vtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.613377 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1181499d-f129-4c48-8376-f01992cc12af-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "1181499d-f129-4c48-8376-f01992cc12af" (UID: "1181499d-f129-4c48-8376-f01992cc12af"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.670067 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2vtx\" (UniqueName: \"kubernetes.io/projected/1181499d-f129-4c48-8376-f01992cc12af-kube-api-access-h2vtx\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:28 crc kubenswrapper[4788]: I1010 16:00:28.670134 4788 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1181499d-f129-4c48-8376-f01992cc12af-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 10 16:00:29 crc kubenswrapper[4788]: I1010 16:00:29.075433 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-4zd6c" Oct 10 16:00:29 crc kubenswrapper[4788]: I1010 16:00:29.079263 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-4zd6c" event={"ID":"1181499d-f129-4c48-8376-f01992cc12af","Type":"ContainerDied","Data":"e949139ebbdc6105b26b7139d59090c145971b1f5766457a2fcc982e6915a19b"} Oct 10 16:00:29 crc kubenswrapper[4788]: I1010 16:00:29.079337 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e949139ebbdc6105b26b7139d59090c145971b1f5766457a2fcc982e6915a19b" Oct 10 16:00:36 crc kubenswrapper[4788]: I1010 16:00:36.233877 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 16:00:37 crc kubenswrapper[4788]: I1010 16:00:37.148196 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"6df1945d6fe1182ef3cdc4c7a6a92c339bc56228bcd7417f29306d9a6b290161"} Oct 10 16:01:11 crc kubenswrapper[4788]: I1010 16:01:11.648269 4788 scope.go:117] "RemoveContainer" containerID="15278051e22f4349d19272666fe9aef820f9cbc1325052282f79fa116cecb9a4" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.661568 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6nm9r"] Oct 10 16:02:00 crc kubenswrapper[4788]: E1010 16:02:00.663045 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1181499d-f129-4c48-8376-f01992cc12af" containerName="storage" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.663064 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1181499d-f129-4c48-8376-f01992cc12af" containerName="storage" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.663483 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1181499d-f129-4c48-8376-f01992cc12af" containerName="storage" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.668741 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.699266 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6nm9r"] Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.808661 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-catalog-content\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.808730 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfjcn\" (UniqueName: \"kubernetes.io/projected/582c358b-0330-4157-bf4c-a95f9d2f30fe-kube-api-access-nfjcn\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.808840 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-utilities\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.910050 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-catalog-content\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.910103 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfjcn\" (UniqueName: \"kubernetes.io/projected/582c358b-0330-4157-bf4c-a95f9d2f30fe-kube-api-access-nfjcn\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.910178 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-utilities\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.910750 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-catalog-content\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.910762 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-utilities\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.934991 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfjcn\" (UniqueName: \"kubernetes.io/projected/582c358b-0330-4157-bf4c-a95f9d2f30fe-kube-api-access-nfjcn\") pod \"certified-operators-6nm9r\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:00 crc kubenswrapper[4788]: I1010 16:02:00.997484 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:01 crc kubenswrapper[4788]: I1010 16:02:01.283365 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6nm9r"] Oct 10 16:02:01 crc kubenswrapper[4788]: I1010 16:02:01.934520 4788 generic.go:334] "Generic (PLEG): container finished" podID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerID="9ab9cf4dcb94528b62de1aa4cb6d4ac442122a03ffc7a038894e6740d1ac9135" exitCode=0 Oct 10 16:02:01 crc kubenswrapper[4788]: I1010 16:02:01.934581 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nm9r" event={"ID":"582c358b-0330-4157-bf4c-a95f9d2f30fe","Type":"ContainerDied","Data":"9ab9cf4dcb94528b62de1aa4cb6d4ac442122a03ffc7a038894e6740d1ac9135"} Oct 10 16:02:01 crc kubenswrapper[4788]: I1010 16:02:01.934615 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nm9r" event={"ID":"582c358b-0330-4157-bf4c-a95f9d2f30fe","Type":"ContainerStarted","Data":"d9fe526680110540f27358664d1c50f05a5cb76c73ee044c79e8daad14dbe619"} Oct 10 16:02:05 crc kubenswrapper[4788]: I1010 16:02:05.968802 4788 generic.go:334] "Generic (PLEG): container finished" podID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerID="dd0beab431b27abe49d7b4e8daf1d9a90bba99d71d65d4705da11bea0797c614" exitCode=0 Oct 10 16:02:05 crc kubenswrapper[4788]: I1010 16:02:05.968893 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nm9r" event={"ID":"582c358b-0330-4157-bf4c-a95f9d2f30fe","Type":"ContainerDied","Data":"dd0beab431b27abe49d7b4e8daf1d9a90bba99d71d65d4705da11bea0797c614"} Oct 10 16:02:07 crc kubenswrapper[4788]: I1010 16:02:07.994128 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nm9r" event={"ID":"582c358b-0330-4157-bf4c-a95f9d2f30fe","Type":"ContainerStarted","Data":"55ae2ee72bf11735cd34b6d37141e8eaded13fd074e197811668f0073ea3b790"} Oct 10 16:02:08 crc kubenswrapper[4788]: I1010 16:02:08.025381 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6nm9r" podStartSLOduration=2.467359023 podStartE2EDuration="8.025359047s" podCreationTimestamp="2025-10-10 16:02:00 +0000 UTC" firstStartedPulling="2025-10-10 16:02:01.937261852 +0000 UTC m=+4624.386977410" lastFinishedPulling="2025-10-10 16:02:07.495261836 +0000 UTC m=+4629.944977434" observedRunningTime="2025-10-10 16:02:08.016791519 +0000 UTC m=+4630.466507117" watchObservedRunningTime="2025-10-10 16:02:08.025359047 +0000 UTC m=+4630.475074595" Oct 10 16:02:10 crc kubenswrapper[4788]: I1010 16:02:10.997900 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:10 crc kubenswrapper[4788]: I1010 16:02:10.998550 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:11 crc kubenswrapper[4788]: I1010 16:02:11.052841 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.133328 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zd2hf"] Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.137822 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.147228 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zd2hf"] Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.310936 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-catalog-content\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.311498 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5rdj\" (UniqueName: \"kubernetes.io/projected/a8af8729-5fca-4920-b8a1-2ea219328322-kube-api-access-s5rdj\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.311737 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-utilities\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.413559 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-catalog-content\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.413937 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5rdj\" (UniqueName: \"kubernetes.io/projected/a8af8729-5fca-4920-b8a1-2ea219328322-kube-api-access-s5rdj\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.414079 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-utilities\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.414209 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-catalog-content\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.414702 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-utilities\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.442038 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5rdj\" (UniqueName: \"kubernetes.io/projected/a8af8729-5fca-4920-b8a1-2ea219328322-kube-api-access-s5rdj\") pod \"redhat-operators-zd2hf\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:19 crc kubenswrapper[4788]: I1010 16:02:19.475010 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:20 crc kubenswrapper[4788]: I1010 16:02:20.001589 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zd2hf"] Oct 10 16:02:20 crc kubenswrapper[4788]: I1010 16:02:20.108116 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd2hf" event={"ID":"a8af8729-5fca-4920-b8a1-2ea219328322","Type":"ContainerStarted","Data":"ebc901245ee70f7e3e38ad88bfdf3dc4cbafafc6c75acdfa6707fe0252261250"} Oct 10 16:02:21 crc kubenswrapper[4788]: I1010 16:02:21.037697 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 16:02:21 crc kubenswrapper[4788]: I1010 16:02:21.117912 4788 generic.go:334] "Generic (PLEG): container finished" podID="a8af8729-5fca-4920-b8a1-2ea219328322" containerID="f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3" exitCode=0 Oct 10 16:02:21 crc kubenswrapper[4788]: I1010 16:02:21.117969 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd2hf" event={"ID":"a8af8729-5fca-4920-b8a1-2ea219328322","Type":"ContainerDied","Data":"f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3"} Oct 10 16:02:22 crc kubenswrapper[4788]: I1010 16:02:22.127175 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd2hf" event={"ID":"a8af8729-5fca-4920-b8a1-2ea219328322","Type":"ContainerStarted","Data":"f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097"} Oct 10 16:02:22 crc kubenswrapper[4788]: I1010 16:02:22.984338 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6nm9r"] Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.134746 4788 generic.go:334] "Generic (PLEG): container finished" podID="a8af8729-5fca-4920-b8a1-2ea219328322" containerID="f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097" exitCode=0 Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.134790 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd2hf" event={"ID":"a8af8729-5fca-4920-b8a1-2ea219328322","Type":"ContainerDied","Data":"f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097"} Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.302022 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lz6z2"] Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.302507 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lz6z2" podUID="f4845d06-1130-4527-a881-c01fbec64e87" containerName="registry-server" containerID="cri-o://927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252" gracePeriod=2 Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.779714 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.844687 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-catalog-content\") pod \"f4845d06-1130-4527-a881-c01fbec64e87\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.844776 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-utilities\") pod \"f4845d06-1130-4527-a881-c01fbec64e87\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.844965 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcsw7\" (UniqueName: \"kubernetes.io/projected/f4845d06-1130-4527-a881-c01fbec64e87-kube-api-access-bcsw7\") pod \"f4845d06-1130-4527-a881-c01fbec64e87\" (UID: \"f4845d06-1130-4527-a881-c01fbec64e87\") " Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.853570 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-utilities" (OuterVolumeSpecName: "utilities") pod "f4845d06-1130-4527-a881-c01fbec64e87" (UID: "f4845d06-1130-4527-a881-c01fbec64e87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.855508 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4845d06-1130-4527-a881-c01fbec64e87-kube-api-access-bcsw7" (OuterVolumeSpecName: "kube-api-access-bcsw7") pod "f4845d06-1130-4527-a881-c01fbec64e87" (UID: "f4845d06-1130-4527-a881-c01fbec64e87"). InnerVolumeSpecName "kube-api-access-bcsw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.903444 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4845d06-1130-4527-a881-c01fbec64e87" (UID: "f4845d06-1130-4527-a881-c01fbec64e87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.946456 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcsw7\" (UniqueName: \"kubernetes.io/projected/f4845d06-1130-4527-a881-c01fbec64e87-kube-api-access-bcsw7\") on node \"crc\" DevicePath \"\"" Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.946486 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:02:23 crc kubenswrapper[4788]: I1010 16:02:23.946496 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4845d06-1130-4527-a881-c01fbec64e87-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.144045 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd2hf" event={"ID":"a8af8729-5fca-4920-b8a1-2ea219328322","Type":"ContainerStarted","Data":"1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741"} Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.148452 4788 generic.go:334] "Generic (PLEG): container finished" podID="f4845d06-1130-4527-a881-c01fbec64e87" containerID="927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252" exitCode=0 Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.148507 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz6z2" event={"ID":"f4845d06-1130-4527-a881-c01fbec64e87","Type":"ContainerDied","Data":"927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252"} Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.148529 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz6z2" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.148552 4788 scope.go:117] "RemoveContainer" containerID="927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.148538 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz6z2" event={"ID":"f4845d06-1130-4527-a881-c01fbec64e87","Type":"ContainerDied","Data":"ada4a0ead8ff1ef4be4a486bd0be518aa493a05c55b05e5da85f59321db08782"} Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.186808 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zd2hf" podStartSLOduration=2.732421865 podStartE2EDuration="5.186782775s" podCreationTimestamp="2025-10-10 16:02:19 +0000 UTC" firstStartedPulling="2025-10-10 16:02:21.12003928 +0000 UTC m=+4643.569754828" lastFinishedPulling="2025-10-10 16:02:23.57440019 +0000 UTC m=+4646.024115738" observedRunningTime="2025-10-10 16:02:24.185018008 +0000 UTC m=+4646.634733576" watchObservedRunningTime="2025-10-10 16:02:24.186782775 +0000 UTC m=+4646.636498333" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.195160 4788 scope.go:117] "RemoveContainer" containerID="06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.226339 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lz6z2"] Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.229826 4788 scope.go:117] "RemoveContainer" containerID="ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.247852 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lz6z2"] Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.262335 4788 scope.go:117] "RemoveContainer" containerID="927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252" Oct 10 16:02:24 crc kubenswrapper[4788]: E1010 16:02:24.265566 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252\": container with ID starting with 927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252 not found: ID does not exist" containerID="927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.265619 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252"} err="failed to get container status \"927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252\": rpc error: code = NotFound desc = could not find container \"927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252\": container with ID starting with 927f1ed8124708621781ba10b7ed4341a84b0acd12e0afe8a6d6e8b0553b6252 not found: ID does not exist" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.265651 4788 scope.go:117] "RemoveContainer" containerID="06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0" Oct 10 16:02:24 crc kubenswrapper[4788]: E1010 16:02:24.267064 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0\": container with ID starting with 06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0 not found: ID does not exist" containerID="06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.267103 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0"} err="failed to get container status \"06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0\": rpc error: code = NotFound desc = could not find container \"06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0\": container with ID starting with 06c105cf528eee1cb09b7d92e0739e1dd893daa0791b7556c924fa58ee05d5b0 not found: ID does not exist" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.267130 4788 scope.go:117] "RemoveContainer" containerID="ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489" Oct 10 16:02:24 crc kubenswrapper[4788]: E1010 16:02:24.267411 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489\": container with ID starting with ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489 not found: ID does not exist" containerID="ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489" Oct 10 16:02:24 crc kubenswrapper[4788]: I1010 16:02:24.267436 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489"} err="failed to get container status \"ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489\": rpc error: code = NotFound desc = could not find container \"ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489\": container with ID starting with ff8f6182a2611007882581dffe12d0f41a1071755df74518ab07f2cd27177489 not found: ID does not exist" Oct 10 16:02:26 crc kubenswrapper[4788]: I1010 16:02:26.249601 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4845d06-1130-4527-a881-c01fbec64e87" path="/var/lib/kubelet/pods/f4845d06-1130-4527-a881-c01fbec64e87/volumes" Oct 10 16:02:29 crc kubenswrapper[4788]: I1010 16:02:29.475903 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:29 crc kubenswrapper[4788]: I1010 16:02:29.476408 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:29 crc kubenswrapper[4788]: I1010 16:02:29.558791 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:30 crc kubenswrapper[4788]: I1010 16:02:30.246761 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:31 crc kubenswrapper[4788]: I1010 16:02:31.700902 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zd2hf"] Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.213225 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zd2hf" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" containerName="registry-server" containerID="cri-o://1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741" gracePeriod=2 Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.745515 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.798048 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5rdj\" (UniqueName: \"kubernetes.io/projected/a8af8729-5fca-4920-b8a1-2ea219328322-kube-api-access-s5rdj\") pod \"a8af8729-5fca-4920-b8a1-2ea219328322\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.798326 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-utilities\") pod \"a8af8729-5fca-4920-b8a1-2ea219328322\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.798433 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-catalog-content\") pod \"a8af8729-5fca-4920-b8a1-2ea219328322\" (UID: \"a8af8729-5fca-4920-b8a1-2ea219328322\") " Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.801118 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-utilities" (OuterVolumeSpecName: "utilities") pod "a8af8729-5fca-4920-b8a1-2ea219328322" (UID: "a8af8729-5fca-4920-b8a1-2ea219328322"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.809134 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8af8729-5fca-4920-b8a1-2ea219328322-kube-api-access-s5rdj" (OuterVolumeSpecName: "kube-api-access-s5rdj") pod "a8af8729-5fca-4920-b8a1-2ea219328322" (UID: "a8af8729-5fca-4920-b8a1-2ea219328322"). InnerVolumeSpecName "kube-api-access-s5rdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.899933 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5rdj\" (UniqueName: \"kubernetes.io/projected/a8af8729-5fca-4920-b8a1-2ea219328322-kube-api-access-s5rdj\") on node \"crc\" DevicePath \"\"" Oct 10 16:02:32 crc kubenswrapper[4788]: I1010 16:02:32.899970 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.225737 4788 generic.go:334] "Generic (PLEG): container finished" podID="a8af8729-5fca-4920-b8a1-2ea219328322" containerID="1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741" exitCode=0 Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.225892 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zd2hf" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.225889 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd2hf" event={"ID":"a8af8729-5fca-4920-b8a1-2ea219328322","Type":"ContainerDied","Data":"1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741"} Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.226483 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zd2hf" event={"ID":"a8af8729-5fca-4920-b8a1-2ea219328322","Type":"ContainerDied","Data":"ebc901245ee70f7e3e38ad88bfdf3dc4cbafafc6c75acdfa6707fe0252261250"} Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.226518 4788 scope.go:117] "RemoveContainer" containerID="1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.263955 4788 scope.go:117] "RemoveContainer" containerID="f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.284052 4788 scope.go:117] "RemoveContainer" containerID="f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.327984 4788 scope.go:117] "RemoveContainer" containerID="1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741" Oct 10 16:02:33 crc kubenswrapper[4788]: E1010 16:02:33.328616 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741\": container with ID starting with 1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741 not found: ID does not exist" containerID="1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.328680 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741"} err="failed to get container status \"1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741\": rpc error: code = NotFound desc = could not find container \"1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741\": container with ID starting with 1e1d58cc11795033ef3a9e2b1ab539e89e07e92d8f085687bb9380643bd49741 not found: ID does not exist" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.328727 4788 scope.go:117] "RemoveContainer" containerID="f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097" Oct 10 16:02:33 crc kubenswrapper[4788]: E1010 16:02:33.329338 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097\": container with ID starting with f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097 not found: ID does not exist" containerID="f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.329390 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097"} err="failed to get container status \"f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097\": rpc error: code = NotFound desc = could not find container \"f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097\": container with ID starting with f5b997f59c8910a35f54e26afa6bf13df44ed448e3cb8a1c4df9306f52032097 not found: ID does not exist" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.329477 4788 scope.go:117] "RemoveContainer" containerID="f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3" Oct 10 16:02:33 crc kubenswrapper[4788]: E1010 16:02:33.330056 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3\": container with ID starting with f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3 not found: ID does not exist" containerID="f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.330092 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3"} err="failed to get container status \"f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3\": rpc error: code = NotFound desc = could not find container \"f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3\": container with ID starting with f9d50d0b22a35910971b5addfc8e2f2f6c9f33993f01bb8b4f4df1349b7969b3 not found: ID does not exist" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.840868 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8af8729-5fca-4920-b8a1-2ea219328322" (UID: "a8af8729-5fca-4920-b8a1-2ea219328322"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:02:33 crc kubenswrapper[4788]: I1010 16:02:33.919430 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8af8729-5fca-4920-b8a1-2ea219328322-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:02:34 crc kubenswrapper[4788]: I1010 16:02:34.174441 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zd2hf"] Oct 10 16:02:34 crc kubenswrapper[4788]: I1010 16:02:34.178460 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zd2hf"] Oct 10 16:02:34 crc kubenswrapper[4788]: I1010 16:02:34.250108 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" path="/var/lib/kubelet/pods/a8af8729-5fca-4920-b8a1-2ea219328322/volumes" Oct 10 16:02:59 crc kubenswrapper[4788]: I1010 16:02:59.406509 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:02:59 crc kubenswrapper[4788]: I1010 16:02:59.407519 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:03:29 crc kubenswrapper[4788]: I1010 16:03:29.405875 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:03:29 crc kubenswrapper[4788]: I1010 16:03:29.407129 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.766515 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-bcwth"] Oct 10 16:03:37 crc kubenswrapper[4788]: E1010 16:03:37.767577 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" containerName="extract-utilities" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.767593 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" containerName="extract-utilities" Oct 10 16:03:37 crc kubenswrapper[4788]: E1010 16:03:37.767608 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" containerName="extract-content" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.767614 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" containerName="extract-content" Oct 10 16:03:37 crc kubenswrapper[4788]: E1010 16:03:37.767620 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" containerName="registry-server" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.767626 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" containerName="registry-server" Oct 10 16:03:37 crc kubenswrapper[4788]: E1010 16:03:37.767638 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4845d06-1130-4527-a881-c01fbec64e87" containerName="extract-content" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.767644 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4845d06-1130-4527-a881-c01fbec64e87" containerName="extract-content" Oct 10 16:03:37 crc kubenswrapper[4788]: E1010 16:03:37.767668 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4845d06-1130-4527-a881-c01fbec64e87" containerName="registry-server" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.767674 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4845d06-1130-4527-a881-c01fbec64e87" containerName="registry-server" Oct 10 16:03:37 crc kubenswrapper[4788]: E1010 16:03:37.767687 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4845d06-1130-4527-a881-c01fbec64e87" containerName="extract-utilities" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.767692 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4845d06-1130-4527-a881-c01fbec64e87" containerName="extract-utilities" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.767842 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8af8729-5fca-4920-b8a1-2ea219328322" containerName="registry-server" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.767854 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4845d06-1130-4527-a881-c01fbec64e87" containerName="registry-server" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.768761 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.771241 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.771578 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.771719 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.772628 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8xgw6" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.774023 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.783463 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-bcwth"] Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.876768 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.877084 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tkdj\" (UniqueName: \"kubernetes.io/projected/2affd330-dfae-44c4-bf13-57e0808e0919-kube-api-access-5tkdj\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.877204 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-config\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.978908 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.978974 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tkdj\" (UniqueName: \"kubernetes.io/projected/2affd330-dfae-44c4-bf13-57e0808e0919-kube-api-access-5tkdj\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.978995 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-config\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.979885 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-config\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:37 crc kubenswrapper[4788]: I1010 16:03:37.980590 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.006044 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tkdj\" (UniqueName: \"kubernetes.io/projected/2affd330-dfae-44c4-bf13-57e0808e0919-kube-api-access-5tkdj\") pod \"dnsmasq-dns-5d7b5456f5-bcwth\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.063011 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-4kbfp"] Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.064498 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.084261 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-4kbfp"] Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.090257 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.182928 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.182982 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpf6p\" (UniqueName: \"kubernetes.io/projected/7f13bc87-6c3a-4620-9a08-d7c9774e7951-kube-api-access-gpf6p\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.183044 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-config\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.284714 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-config\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.285260 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.285295 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpf6p\" (UniqueName: \"kubernetes.io/projected/7f13bc87-6c3a-4620-9a08-d7c9774e7951-kube-api-access-gpf6p\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.286350 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-config\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.286937 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.306296 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpf6p\" (UniqueName: \"kubernetes.io/projected/7f13bc87-6c3a-4620-9a08-d7c9774e7951-kube-api-access-gpf6p\") pod \"dnsmasq-dns-98ddfc8f-4kbfp\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.390605 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.610394 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-4kbfp"] Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.639345 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-bcwth"] Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.890566 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.891757 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.896222 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.896333 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.896531 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.896626 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-jp2xm" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.896915 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.918665 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.999358 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.999462 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tqgm\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-kube-api-access-5tqgm\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.999493 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.999553 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.999667 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.999717 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:38 crc kubenswrapper[4788]: I1010 16:03:38.999761 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:38.999944 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:38.999983 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: W1010 16:03:39.007713 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2affd330_dfae_44c4_bf13_57e0808e0919.slice/crio-3bff89a7293b47e9efdccf607ad55124e4d4ee1e06192d0ea0dffd262dfdfa1f WatchSource:0}: Error finding container 3bff89a7293b47e9efdccf607ad55124e4d4ee1e06192d0ea0dffd262dfdfa1f: Status 404 returned error can't find the container with id 3bff89a7293b47e9efdccf607ad55124e4d4ee1e06192d0ea0dffd262dfdfa1f Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101132 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101203 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101236 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101288 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tqgm\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-kube-api-access-5tqgm\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101310 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101338 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101359 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101378 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.101397 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.102514 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.102963 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.102872 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.108287 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.110968 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.111059 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/499672cb3e8c066ee515fd533ccb139a382220336e9a03a35c83880afd837990/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.112330 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.114005 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.115232 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.122419 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tqgm\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-kube-api-access-5tqgm\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.153438 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"rabbitmq-server-0\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.213522 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.238563 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.239936 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.242478 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.242847 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.242963 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.243979 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qcsmp" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.246084 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.266927 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306106 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306234 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075856c-3082-4eee-99fc-d283a5ee1a93-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306292 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306307 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075856c-3082-4eee-99fc-d283a5ee1a93-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306339 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306359 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306554 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgpfk\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-kube-api-access-qgpfk\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306720 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.306909 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409227 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409293 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075856c-3082-4eee-99fc-d283a5ee1a93-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409345 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409370 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409409 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgpfk\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-kube-api-access-qgpfk\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409433 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409484 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409538 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.409581 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075856c-3082-4eee-99fc-d283a5ee1a93-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.415406 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.416095 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.416267 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075856c-3082-4eee-99fc-d283a5ee1a93-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.417758 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.418555 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.418585 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f86df4606710da878b6ae934207ef32850fa01ed5a0881365f46077ad27e410e/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.419076 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.427893 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075856c-3082-4eee-99fc-d283a5ee1a93-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.429642 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.432519 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgpfk\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-kube-api-access-qgpfk\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.460231 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.574057 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.710175 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.884085 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.885936 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.893751 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.894412 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.894586 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.894861 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.894974 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-x86bm" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.896201 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba","Type":"ContainerStarted","Data":"fcae49447a8f002f1eabfea45c0bb6db2b1baf3fd85f0c3e2b9bba2554c3fa3d"} Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.899600 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.900674 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.903136 4788 generic.go:334] "Generic (PLEG): container finished" podID="2affd330-dfae-44c4-bf13-57e0808e0919" containerID="568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5" exitCode=0 Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.903252 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" event={"ID":"2affd330-dfae-44c4-bf13-57e0808e0919","Type":"ContainerDied","Data":"568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5"} Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.903305 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" event={"ID":"2affd330-dfae-44c4-bf13-57e0808e0919","Type":"ContainerStarted","Data":"3bff89a7293b47e9efdccf607ad55124e4d4ee1e06192d0ea0dffd262dfdfa1f"} Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.908261 4788 generic.go:334] "Generic (PLEG): container finished" podID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" containerID="812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee" exitCode=0 Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.908310 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" event={"ID":"7f13bc87-6c3a-4620-9a08-d7c9774e7951","Type":"ContainerDied","Data":"812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee"} Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.908344 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" event={"ID":"7f13bc87-6c3a-4620-9a08-d7c9774e7951","Type":"ContainerStarted","Data":"523b2a05e9cd7a2621c98b826756a69e174ccc685e27e33650bf0223ddb0002b"} Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.917842 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3b719e3b-dfdb-414c-9618-be4914132848\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b719e3b-dfdb-414c-9618-be4914132848\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.918054 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.918088 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1690fbb-be06-482f-9c94-4d436d9b73b1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.918120 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.918157 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-secrets\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.918181 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.918212 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.918241 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: I1010 16:03:39.918270 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk7pk\" (UniqueName: \"kubernetes.io/projected/f1690fbb-be06-482f-9c94-4d436d9b73b1-kube-api-access-xk7pk\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:39 crc kubenswrapper[4788]: E1010 16:03:39.923799 4788 mount_linux.go:282] Mount failed: exit status 32 Oct 10 16:03:39 crc kubenswrapper[4788]: Mounting command: mount Oct 10 16:03:39 crc kubenswrapper[4788]: Mounting arguments: --no-canonicalize -o bind /proc/4788/fd/24 /var/lib/kubelet/pods/2affd330-dfae-44c4-bf13-57e0808e0919/volume-subpaths/dns-svc/dnsmasq-dns/1 Oct 10 16:03:39 crc kubenswrapper[4788]: Output: mount: /var/lib/kubelet/pods/2affd330-dfae-44c4-bf13-57e0808e0919/volume-subpaths/dns-svc/dnsmasq-dns/1: mount(2) system call failed: No such file or directory. Oct 10 16:03:39 crc kubenswrapper[4788]: E1010 16:03:39.943577 4788 kubelet_pods.go:349] "Failed to prepare subPath for volumeMount of the container" err=< Oct 10 16:03:39 crc kubenswrapper[4788]: error mounting /var/lib/kubelet/pods/2affd330-dfae-44c4-bf13-57e0808e0919/volumes/kubernetes.io~configmap/dns-svc/..2025_10_10_16_03_37.1694214499/dns-svc: mount failed: exit status 32 Oct 10 16:03:39 crc kubenswrapper[4788]: Mounting command: mount Oct 10 16:03:39 crc kubenswrapper[4788]: Mounting arguments: --no-canonicalize -o bind /proc/4788/fd/24 /var/lib/kubelet/pods/2affd330-dfae-44c4-bf13-57e0808e0919/volume-subpaths/dns-svc/dnsmasq-dns/1 Oct 10 16:03:39 crc kubenswrapper[4788]: Output: mount: /var/lib/kubelet/pods/2affd330-dfae-44c4-bf13-57e0808e0919/volume-subpaths/dns-svc/dnsmasq-dns/1: mount(2) system call failed: No such file or directory. Oct 10 16:03:39 crc kubenswrapper[4788]: > containerName="dnsmasq-dns" volumeMountName="dns-svc" Oct 10 16:03:39 crc kubenswrapper[4788]: E1010 16:03:39.943804 4788 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5tkdj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d7b5456f5-bcwth_openstack(2affd330-dfae-44c4-bf13-57e0808e0919): CreateContainerConfigError: failed to prepare subPath for volumeMount \"dns-svc\" of container \"dnsmasq-dns\"" logger="UnhandledError" Oct 10 16:03:39 crc kubenswrapper[4788]: E1010 16:03:39.945649 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerConfigError: \"failed to prepare subPath for volumeMount \\\"dns-svc\\\" of container \\\"dnsmasq-dns\\\"\"" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" podUID="2affd330-dfae-44c4-bf13-57e0808e0919" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.023678 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-secrets\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.023740 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.023786 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.023829 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.023859 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk7pk\" (UniqueName: \"kubernetes.io/projected/f1690fbb-be06-482f-9c94-4d436d9b73b1-kube-api-access-xk7pk\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.023936 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3b719e3b-dfdb-414c-9618-be4914132848\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b719e3b-dfdb-414c-9618-be4914132848\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.024041 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.024068 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1690fbb-be06-482f-9c94-4d436d9b73b1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.024091 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.026989 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.028834 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.029708 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1690fbb-be06-482f-9c94-4d436d9b73b1-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.029853 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-secrets\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.029971 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.031638 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1690fbb-be06-482f-9c94-4d436d9b73b1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.032973 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.033370 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3b719e3b-dfdb-414c-9618-be4914132848\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b719e3b-dfdb-414c-9618-be4914132848\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1bc07e92fc9f9569bc8bcdbe4ea213c8f822e0c9913e022d7351e5fce557599f/globalmount\"" pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.039422 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1690fbb-be06-482f-9c94-4d436d9b73b1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.050940 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk7pk\" (UniqueName: \"kubernetes.io/projected/f1690fbb-be06-482f-9c94-4d436d9b73b1-kube-api-access-xk7pk\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.065380 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:03:40 crc kubenswrapper[4788]: W1010 16:03:40.077316 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3075856c_3082_4eee_99fc_d283a5ee1a93.slice/crio-8d81c74625572dcdacee5230e950a6a48c9996de6350c771d688198a655161aa WatchSource:0}: Error finding container 8d81c74625572dcdacee5230e950a6a48c9996de6350c771d688198a655161aa: Status 404 returned error can't find the container with id 8d81c74625572dcdacee5230e950a6a48c9996de6350c771d688198a655161aa Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.089180 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3b719e3b-dfdb-414c-9618-be4914132848\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b719e3b-dfdb-414c-9618-be4914132848\") pod \"openstack-galera-0\" (UID: \"f1690fbb-be06-482f-9c94-4d436d9b73b1\") " pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.215787 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.394420 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.396160 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.398813 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.401214 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-9wm8m" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.423336 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.431847 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzvmv\" (UniqueName: \"kubernetes.io/projected/230985d2-0cc5-499d-ad3a-dd37be0b450b-kube-api-access-pzvmv\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.431887 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/230985d2-0cc5-499d-ad3a-dd37be0b450b-config-data\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.431926 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/230985d2-0cc5-499d-ad3a-dd37be0b450b-kolla-config\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.534619 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzvmv\" (UniqueName: \"kubernetes.io/projected/230985d2-0cc5-499d-ad3a-dd37be0b450b-kube-api-access-pzvmv\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.534693 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/230985d2-0cc5-499d-ad3a-dd37be0b450b-config-data\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.534736 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/230985d2-0cc5-499d-ad3a-dd37be0b450b-kolla-config\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.536772 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/230985d2-0cc5-499d-ad3a-dd37be0b450b-config-data\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.536921 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/230985d2-0cc5-499d-ad3a-dd37be0b450b-kolla-config\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.546823 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.555776 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzvmv\" (UniqueName: \"kubernetes.io/projected/230985d2-0cc5-499d-ad3a-dd37be0b450b-kube-api-access-pzvmv\") pod \"memcached-0\" (UID: \"230985d2-0cc5-499d-ad3a-dd37be0b450b\") " pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.728461 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.946627 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f1690fbb-be06-482f-9c94-4d436d9b73b1","Type":"ContainerStarted","Data":"bf2bd4da0f5376ecb18ddabf3b2f4dbce82ef3afab50a8a10154d9d209196f54"} Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.947156 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f1690fbb-be06-482f-9c94-4d436d9b73b1","Type":"ContainerStarted","Data":"f12ee60b2a7e174dfb2a4eec722f3895490b94081db20fe1ad50b415bea762dc"} Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.950550 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" event={"ID":"7f13bc87-6c3a-4620-9a08-d7c9774e7951","Type":"ContainerStarted","Data":"2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead"} Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.950747 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:40 crc kubenswrapper[4788]: I1010 16:03:40.953699 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075856c-3082-4eee-99fc-d283a5ee1a93","Type":"ContainerStarted","Data":"8d81c74625572dcdacee5230e950a6a48c9996de6350c771d688198a655161aa"} Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.000241 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" podStartSLOduration=3.000216947 podStartE2EDuration="3.000216947s" podCreationTimestamp="2025-10-10 16:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:03:40.996981711 +0000 UTC m=+4723.446697259" watchObservedRunningTime="2025-10-10 16:03:41.000216947 +0000 UTC m=+4723.449932495" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.112035 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.449834 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.452292 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.455315 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-tbr6r" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.456488 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.457125 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.458722 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.460053 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563472 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563553 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563593 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l28xw\" (UniqueName: \"kubernetes.io/projected/a22008ee-3a48-4a02-9301-506b7133a8a5-kube-api-access-l28xw\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563623 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6bbf53ac-a614-4d73-91da-40ebf8b6f42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bbf53ac-a614-4d73-91da-40ebf8b6f42b\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563657 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563676 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563719 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563740 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a22008ee-3a48-4a02-9301-506b7133a8a5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.563762 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.665292 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.665864 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.665916 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l28xw\" (UniqueName: \"kubernetes.io/projected/a22008ee-3a48-4a02-9301-506b7133a8a5-kube-api-access-l28xw\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.665970 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6bbf53ac-a614-4d73-91da-40ebf8b6f42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bbf53ac-a614-4d73-91da-40ebf8b6f42b\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.666019 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.666053 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.666120 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.666195 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a22008ee-3a48-4a02-9301-506b7133a8a5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.666232 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.667707 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.667924 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a22008ee-3a48-4a02-9301-506b7133a8a5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.668307 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.669004 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a22008ee-3a48-4a02-9301-506b7133a8a5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.672773 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.672846 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6bbf53ac-a614-4d73-91da-40ebf8b6f42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bbf53ac-a614-4d73-91da-40ebf8b6f42b\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/720d5e3636df74879a7745ac1cd135e4d8f4786233f26ec24bc0781b0c32911f/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.675159 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.675212 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.676574 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22008ee-3a48-4a02-9301-506b7133a8a5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.694425 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l28xw\" (UniqueName: \"kubernetes.io/projected/a22008ee-3a48-4a02-9301-506b7133a8a5-kube-api-access-l28xw\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.811778 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6bbf53ac-a614-4d73-91da-40ebf8b6f42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6bbf53ac-a614-4d73-91da-40ebf8b6f42b\") pod \"openstack-cell1-galera-0\" (UID: \"a22008ee-3a48-4a02-9301-506b7133a8a5\") " pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.965208 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" event={"ID":"2affd330-dfae-44c4-bf13-57e0808e0919","Type":"ContainerStarted","Data":"2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997"} Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.965490 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.967761 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075856c-3082-4eee-99fc-d283a5ee1a93","Type":"ContainerStarted","Data":"6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b"} Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.972499 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"230985d2-0cc5-499d-ad3a-dd37be0b450b","Type":"ContainerStarted","Data":"4fa4aac8c25de7f9956f90a0c21b1da14b0a2fb6b6533afeaf4b3858534daab1"} Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.972727 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"230985d2-0cc5-499d-ad3a-dd37be0b450b","Type":"ContainerStarted","Data":"fb05e6c678df91b561d615a5719a101670fdeaaaf2dd577420b7359bb8a6d1b7"} Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.973365 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.978102 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba","Type":"ContainerStarted","Data":"ae5520da34fbbf08a7e447550b21b3b1b7be9875137e54a637e2e8a18d14a209"} Oct 10 16:03:41 crc kubenswrapper[4788]: I1010 16:03:41.987531 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" podStartSLOduration=4.98750631 podStartE2EDuration="4.98750631s" podCreationTimestamp="2025-10-10 16:03:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:03:41.982365663 +0000 UTC m=+4724.432081221" watchObservedRunningTime="2025-10-10 16:03:41.98750631 +0000 UTC m=+4724.437221868" Oct 10 16:03:42 crc kubenswrapper[4788]: I1010 16:03:42.031057 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.031036976 podStartE2EDuration="2.031036976s" podCreationTimestamp="2025-10-10 16:03:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:03:42.022080698 +0000 UTC m=+4724.471796256" watchObservedRunningTime="2025-10-10 16:03:42.031036976 +0000 UTC m=+4724.480752544" Oct 10 16:03:42 crc kubenswrapper[4788]: I1010 16:03:42.078858 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:42 crc kubenswrapper[4788]: I1010 16:03:42.534811 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 10 16:03:42 crc kubenswrapper[4788]: W1010 16:03:42.544051 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda22008ee_3a48_4a02_9301_506b7133a8a5.slice/crio-f81a5759dd330cb25114dea89caf1c2fc40f374dd0b7394f2138df0f29c42069 WatchSource:0}: Error finding container f81a5759dd330cb25114dea89caf1c2fc40f374dd0b7394f2138df0f29c42069: Status 404 returned error can't find the container with id f81a5759dd330cb25114dea89caf1c2fc40f374dd0b7394f2138df0f29c42069 Oct 10 16:03:42 crc kubenswrapper[4788]: I1010 16:03:42.988485 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a22008ee-3a48-4a02-9301-506b7133a8a5","Type":"ContainerStarted","Data":"8264484e636f149cd5f632ec77111318828ad3a532b6462f847cec7f73b4feb7"} Oct 10 16:03:42 crc kubenswrapper[4788]: I1010 16:03:42.988955 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a22008ee-3a48-4a02-9301-506b7133a8a5","Type":"ContainerStarted","Data":"f81a5759dd330cb25114dea89caf1c2fc40f374dd0b7394f2138df0f29c42069"} Oct 10 16:03:45 crc kubenswrapper[4788]: I1010 16:03:45.006942 4788 generic.go:334] "Generic (PLEG): container finished" podID="f1690fbb-be06-482f-9c94-4d436d9b73b1" containerID="bf2bd4da0f5376ecb18ddabf3b2f4dbce82ef3afab50a8a10154d9d209196f54" exitCode=0 Oct 10 16:03:45 crc kubenswrapper[4788]: I1010 16:03:45.007379 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f1690fbb-be06-482f-9c94-4d436d9b73b1","Type":"ContainerDied","Data":"bf2bd4da0f5376ecb18ddabf3b2f4dbce82ef3afab50a8a10154d9d209196f54"} Oct 10 16:03:46 crc kubenswrapper[4788]: I1010 16:03:46.017131 4788 generic.go:334] "Generic (PLEG): container finished" podID="a22008ee-3a48-4a02-9301-506b7133a8a5" containerID="8264484e636f149cd5f632ec77111318828ad3a532b6462f847cec7f73b4feb7" exitCode=0 Oct 10 16:03:46 crc kubenswrapper[4788]: I1010 16:03:46.017228 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a22008ee-3a48-4a02-9301-506b7133a8a5","Type":"ContainerDied","Data":"8264484e636f149cd5f632ec77111318828ad3a532b6462f847cec7f73b4feb7"} Oct 10 16:03:46 crc kubenswrapper[4788]: I1010 16:03:46.022163 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f1690fbb-be06-482f-9c94-4d436d9b73b1","Type":"ContainerStarted","Data":"5e20213ee43625761335310dfc123d274e3f966dcece05e061d49e4d0de4a054"} Oct 10 16:03:46 crc kubenswrapper[4788]: I1010 16:03:46.089224 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.089193933 podStartE2EDuration="8.089193933s" podCreationTimestamp="2025-10-10 16:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:03:46.077649347 +0000 UTC m=+4728.527364915" watchObservedRunningTime="2025-10-10 16:03:46.089193933 +0000 UTC m=+4728.538909491" Oct 10 16:03:47 crc kubenswrapper[4788]: I1010 16:03:47.034213 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a22008ee-3a48-4a02-9301-506b7133a8a5","Type":"ContainerStarted","Data":"348c5b5a0f18de16401e6e21e9d060edf2c1e49ded6366c8db00fcc2f3f51405"} Oct 10 16:03:47 crc kubenswrapper[4788]: I1010 16:03:47.064353 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.064327643 podStartE2EDuration="7.064327643s" podCreationTimestamp="2025-10-10 16:03:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:03:47.05783052 +0000 UTC m=+4729.507546078" watchObservedRunningTime="2025-10-10 16:03:47.064327643 +0000 UTC m=+4729.514043231" Oct 10 16:03:47 crc kubenswrapper[4788]: E1010 16:03:47.999710 4788 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.129:55414->38.102.83.129:41727: write tcp 38.102.83.129:55414->38.102.83.129:41727: write: broken pipe Oct 10 16:03:48 crc kubenswrapper[4788]: I1010 16:03:48.092295 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:48 crc kubenswrapper[4788]: I1010 16:03:48.393407 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:03:48 crc kubenswrapper[4788]: I1010 16:03:48.458528 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-bcwth"] Oct 10 16:03:49 crc kubenswrapper[4788]: I1010 16:03:49.050921 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" podUID="2affd330-dfae-44c4-bf13-57e0808e0919" containerName="dnsmasq-dns" containerID="cri-o://2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997" gracePeriod=10 Oct 10 16:03:49 crc kubenswrapper[4788]: I1010 16:03:49.900297 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.018177 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-dns-svc\") pod \"2affd330-dfae-44c4-bf13-57e0808e0919\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.018317 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-config\") pod \"2affd330-dfae-44c4-bf13-57e0808e0919\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.018405 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tkdj\" (UniqueName: \"kubernetes.io/projected/2affd330-dfae-44c4-bf13-57e0808e0919-kube-api-access-5tkdj\") pod \"2affd330-dfae-44c4-bf13-57e0808e0919\" (UID: \"2affd330-dfae-44c4-bf13-57e0808e0919\") " Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.039025 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2affd330-dfae-44c4-bf13-57e0808e0919-kube-api-access-5tkdj" (OuterVolumeSpecName: "kube-api-access-5tkdj") pod "2affd330-dfae-44c4-bf13-57e0808e0919" (UID: "2affd330-dfae-44c4-bf13-57e0808e0919"). InnerVolumeSpecName "kube-api-access-5tkdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.064729 4788 generic.go:334] "Generic (PLEG): container finished" podID="2affd330-dfae-44c4-bf13-57e0808e0919" containerID="2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997" exitCode=0 Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.064795 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" event={"ID":"2affd330-dfae-44c4-bf13-57e0808e0919","Type":"ContainerDied","Data":"2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997"} Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.064833 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" event={"ID":"2affd330-dfae-44c4-bf13-57e0808e0919","Type":"ContainerDied","Data":"3bff89a7293b47e9efdccf607ad55124e4d4ee1e06192d0ea0dffd262dfdfa1f"} Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.064858 4788 scope.go:117] "RemoveContainer" containerID="2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.065071 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-bcwth" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.086919 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-config" (OuterVolumeSpecName: "config") pod "2affd330-dfae-44c4-bf13-57e0808e0919" (UID: "2affd330-dfae-44c4-bf13-57e0808e0919"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.090227 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2affd330-dfae-44c4-bf13-57e0808e0919" (UID: "2affd330-dfae-44c4-bf13-57e0808e0919"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.118664 4788 scope.go:117] "RemoveContainer" containerID="568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.119960 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tkdj\" (UniqueName: \"kubernetes.io/projected/2affd330-dfae-44c4-bf13-57e0808e0919-kube-api-access-5tkdj\") on node \"crc\" DevicePath \"\"" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.120003 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.120017 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2affd330-dfae-44c4-bf13-57e0808e0919-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.138741 4788 scope.go:117] "RemoveContainer" containerID="2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997" Oct 10 16:03:50 crc kubenswrapper[4788]: E1010 16:03:50.139129 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997\": container with ID starting with 2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997 not found: ID does not exist" containerID="2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.139189 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997"} err="failed to get container status \"2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997\": rpc error: code = NotFound desc = could not find container \"2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997\": container with ID starting with 2ce11148a6b218ddd550f218cc0d99ea8593576660791ebf1766c79c259bd997 not found: ID does not exist" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.139219 4788 scope.go:117] "RemoveContainer" containerID="568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5" Oct 10 16:03:50 crc kubenswrapper[4788]: E1010 16:03:50.139757 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5\": container with ID starting with 568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5 not found: ID does not exist" containerID="568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.139792 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5"} err="failed to get container status \"568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5\": rpc error: code = NotFound desc = could not find container \"568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5\": container with ID starting with 568606344fe7f6c5739fae11c2e6ebe17dc8123902f4990f02d584aa29c37ca5 not found: ID does not exist" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.216479 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.216524 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.393753 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-bcwth"] Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.406660 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-bcwth"] Oct 10 16:03:50 crc kubenswrapper[4788]: I1010 16:03:50.730478 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 10 16:03:52 crc kubenswrapper[4788]: I1010 16:03:52.079685 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:52 crc kubenswrapper[4788]: I1010 16:03:52.080107 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:52 crc kubenswrapper[4788]: I1010 16:03:52.243057 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2affd330-dfae-44c4-bf13-57e0808e0919" path="/var/lib/kubelet/pods/2affd330-dfae-44c4-bf13-57e0808e0919/volumes" Oct 10 16:03:52 crc kubenswrapper[4788]: I1010 16:03:52.279175 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 10 16:03:52 crc kubenswrapper[4788]: I1010 16:03:52.346032 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 10 16:03:54 crc kubenswrapper[4788]: I1010 16:03:54.151617 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:54 crc kubenswrapper[4788]: I1010 16:03:54.216257 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 10 16:03:59 crc kubenswrapper[4788]: I1010 16:03:59.406868 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:03:59 crc kubenswrapper[4788]: I1010 16:03:59.407654 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:03:59 crc kubenswrapper[4788]: I1010 16:03:59.407728 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:03:59 crc kubenswrapper[4788]: I1010 16:03:59.408788 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6df1945d6fe1182ef3cdc4c7a6a92c339bc56228bcd7417f29306d9a6b290161"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:03:59 crc kubenswrapper[4788]: I1010 16:03:59.408909 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://6df1945d6fe1182ef3cdc4c7a6a92c339bc56228bcd7417f29306d9a6b290161" gracePeriod=600 Oct 10 16:04:00 crc kubenswrapper[4788]: I1010 16:04:00.170429 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="6df1945d6fe1182ef3cdc4c7a6a92c339bc56228bcd7417f29306d9a6b290161" exitCode=0 Oct 10 16:04:00 crc kubenswrapper[4788]: I1010 16:04:00.170511 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"6df1945d6fe1182ef3cdc4c7a6a92c339bc56228bcd7417f29306d9a6b290161"} Oct 10 16:04:00 crc kubenswrapper[4788]: I1010 16:04:00.170996 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e"} Oct 10 16:04:00 crc kubenswrapper[4788]: I1010 16:04:00.171042 4788 scope.go:117] "RemoveContainer" containerID="b284d26d2d39ce1af065cc42ef03351807266c6cc755868ec0b739ff88a0fee7" Oct 10 16:04:15 crc kubenswrapper[4788]: I1010 16:04:15.318551 4788 generic.go:334] "Generic (PLEG): container finished" podID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerID="6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b" exitCode=0 Oct 10 16:04:15 crc kubenswrapper[4788]: I1010 16:04:15.318689 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075856c-3082-4eee-99fc-d283a5ee1a93","Type":"ContainerDied","Data":"6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b"} Oct 10 16:04:15 crc kubenswrapper[4788]: I1010 16:04:15.321244 4788 generic.go:334] "Generic (PLEG): container finished" podID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerID="ae5520da34fbbf08a7e447550b21b3b1b7be9875137e54a637e2e8a18d14a209" exitCode=0 Oct 10 16:04:15 crc kubenswrapper[4788]: I1010 16:04:15.321282 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba","Type":"ContainerDied","Data":"ae5520da34fbbf08a7e447550b21b3b1b7be9875137e54a637e2e8a18d14a209"} Oct 10 16:04:16 crc kubenswrapper[4788]: I1010 16:04:16.330378 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba","Type":"ContainerStarted","Data":"545de9b4d0601f9de022f68921e332ecc4e75a0b4a62bbefe804e3087734469f"} Oct 10 16:04:16 crc kubenswrapper[4788]: I1010 16:04:16.331060 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 10 16:04:16 crc kubenswrapper[4788]: I1010 16:04:16.333241 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075856c-3082-4eee-99fc-d283a5ee1a93","Type":"ContainerStarted","Data":"daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6"} Oct 10 16:04:16 crc kubenswrapper[4788]: I1010 16:04:16.333483 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:16 crc kubenswrapper[4788]: I1010 16:04:16.358629 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.358603961 podStartE2EDuration="39.358603961s" podCreationTimestamp="2025-10-10 16:03:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:04:16.352954711 +0000 UTC m=+4758.802670269" watchObservedRunningTime="2025-10-10 16:04:16.358603961 +0000 UTC m=+4758.808319539" Oct 10 16:04:16 crc kubenswrapper[4788]: I1010 16:04:16.382220 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.382200847 podStartE2EDuration="38.382200847s" podCreationTimestamp="2025-10-10 16:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:04:16.377514582 +0000 UTC m=+4758.827230160" watchObservedRunningTime="2025-10-10 16:04:16.382200847 +0000 UTC m=+4758.831916395" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.477240 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ffbxz"] Oct 10 16:04:22 crc kubenswrapper[4788]: E1010 16:04:22.479188 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2affd330-dfae-44c4-bf13-57e0808e0919" containerName="dnsmasq-dns" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.479289 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2affd330-dfae-44c4-bf13-57e0808e0919" containerName="dnsmasq-dns" Oct 10 16:04:22 crc kubenswrapper[4788]: E1010 16:04:22.479331 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2affd330-dfae-44c4-bf13-57e0808e0919" containerName="init" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.479339 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2affd330-dfae-44c4-bf13-57e0808e0919" containerName="init" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.479724 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="2affd330-dfae-44c4-bf13-57e0808e0919" containerName="dnsmasq-dns" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.481881 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.498309 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ffbxz"] Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.599911 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv8rj\" (UniqueName: \"kubernetes.io/projected/934209ba-9aa9-4ec2-8572-37c7e924ecfe-kube-api-access-cv8rj\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.600003 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-catalog-content\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.600043 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-utilities\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.701591 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv8rj\" (UniqueName: \"kubernetes.io/projected/934209ba-9aa9-4ec2-8572-37c7e924ecfe-kube-api-access-cv8rj\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.702257 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-catalog-content\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.702309 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-utilities\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.702859 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-catalog-content\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.703027 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-utilities\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.735999 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv8rj\" (UniqueName: \"kubernetes.io/projected/934209ba-9aa9-4ec2-8572-37c7e924ecfe-kube-api-access-cv8rj\") pod \"redhat-marketplace-ffbxz\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:22 crc kubenswrapper[4788]: I1010 16:04:22.821548 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:23 crc kubenswrapper[4788]: I1010 16:04:23.341738 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ffbxz"] Oct 10 16:04:23 crc kubenswrapper[4788]: I1010 16:04:23.402949 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ffbxz" event={"ID":"934209ba-9aa9-4ec2-8572-37c7e924ecfe","Type":"ContainerStarted","Data":"f9b8b25d1bfff1c4319d58b0dacf334c9eef44b45e057e7f735cb40d9a889569"} Oct 10 16:04:24 crc kubenswrapper[4788]: I1010 16:04:24.412661 4788 generic.go:334] "Generic (PLEG): container finished" podID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerID="a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595" exitCode=0 Oct 10 16:04:24 crc kubenswrapper[4788]: I1010 16:04:24.412720 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ffbxz" event={"ID":"934209ba-9aa9-4ec2-8572-37c7e924ecfe","Type":"ContainerDied","Data":"a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595"} Oct 10 16:04:26 crc kubenswrapper[4788]: I1010 16:04:26.435095 4788 generic.go:334] "Generic (PLEG): container finished" podID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerID="f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb" exitCode=0 Oct 10 16:04:26 crc kubenswrapper[4788]: I1010 16:04:26.435221 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ffbxz" event={"ID":"934209ba-9aa9-4ec2-8572-37c7e924ecfe","Type":"ContainerDied","Data":"f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb"} Oct 10 16:04:27 crc kubenswrapper[4788]: I1010 16:04:27.450941 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ffbxz" event={"ID":"934209ba-9aa9-4ec2-8572-37c7e924ecfe","Type":"ContainerStarted","Data":"ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8"} Oct 10 16:04:27 crc kubenswrapper[4788]: I1010 16:04:27.473122 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ffbxz" podStartSLOduration=2.938623651 podStartE2EDuration="5.473105559s" podCreationTimestamp="2025-10-10 16:04:22 +0000 UTC" firstStartedPulling="2025-10-10 16:04:24.41488343 +0000 UTC m=+4766.864598988" lastFinishedPulling="2025-10-10 16:04:26.949365348 +0000 UTC m=+4769.399080896" observedRunningTime="2025-10-10 16:04:27.471835625 +0000 UTC m=+4769.921551213" watchObservedRunningTime="2025-10-10 16:04:27.473105559 +0000 UTC m=+4769.922821107" Oct 10 16:04:29 crc kubenswrapper[4788]: I1010 16:04:29.217428 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 10 16:04:29 crc kubenswrapper[4788]: I1010 16:04:29.576313 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:32 crc kubenswrapper[4788]: I1010 16:04:32.821883 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:32 crc kubenswrapper[4788]: I1010 16:04:32.823970 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:32 crc kubenswrapper[4788]: I1010 16:04:32.869165 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:33 crc kubenswrapper[4788]: I1010 16:04:33.548245 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:33 crc kubenswrapper[4788]: I1010 16:04:33.611120 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ffbxz"] Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.381709 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-92hfq"] Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.383085 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.411071 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-92hfq"] Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.554154 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ssw\" (UniqueName: \"kubernetes.io/projected/436a3bba-9371-4bf0-938d-8655a4862e22-kube-api-access-j9ssw\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.554689 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-config\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.554734 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.655859 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.656011 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ssw\" (UniqueName: \"kubernetes.io/projected/436a3bba-9371-4bf0-938d-8655a4862e22-kube-api-access-j9ssw\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.656091 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-config\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.656826 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.657025 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-config\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.675949 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ssw\" (UniqueName: \"kubernetes.io/projected/436a3bba-9371-4bf0-938d-8655a4862e22-kube-api-access-j9ssw\") pod \"dnsmasq-dns-5b7946d7b9-92hfq\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:34 crc kubenswrapper[4788]: I1010 16:04:34.729890 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:35 crc kubenswrapper[4788]: I1010 16:04:35.052517 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:04:35 crc kubenswrapper[4788]: I1010 16:04:35.216226 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-92hfq"] Oct 10 16:04:35 crc kubenswrapper[4788]: W1010 16:04:35.224834 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod436a3bba_9371_4bf0_938d_8655a4862e22.slice/crio-bc7e81f9cc68f3615c84815044e3cf9562cd489a4ef457f8598f903d43ef7abe WatchSource:0}: Error finding container bc7e81f9cc68f3615c84815044e3cf9562cd489a4ef457f8598f903d43ef7abe: Status 404 returned error can't find the container with id bc7e81f9cc68f3615c84815044e3cf9562cd489a4ef457f8598f903d43ef7abe Oct 10 16:04:35 crc kubenswrapper[4788]: I1010 16:04:35.510374 4788 generic.go:334] "Generic (PLEG): container finished" podID="436a3bba-9371-4bf0-938d-8655a4862e22" containerID="cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa" exitCode=0 Oct 10 16:04:35 crc kubenswrapper[4788]: I1010 16:04:35.510924 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ffbxz" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerName="registry-server" containerID="cri-o://ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8" gracePeriod=2 Oct 10 16:04:35 crc kubenswrapper[4788]: I1010 16:04:35.513928 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" event={"ID":"436a3bba-9371-4bf0-938d-8655a4862e22","Type":"ContainerDied","Data":"cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa"} Oct 10 16:04:35 crc kubenswrapper[4788]: I1010 16:04:35.513994 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" event={"ID":"436a3bba-9371-4bf0-938d-8655a4862e22","Type":"ContainerStarted","Data":"bc7e81f9cc68f3615c84815044e3cf9562cd489a4ef457f8598f903d43ef7abe"} Oct 10 16:04:35 crc kubenswrapper[4788]: E1010 16:04:35.613921 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod436a3bba_9371_4bf0_938d_8655a4862e22.slice/crio-cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa.scope\": RecentStats: unable to find data in memory cache]" Oct 10 16:04:35 crc kubenswrapper[4788]: I1010 16:04:35.871082 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:04:35 crc kubenswrapper[4788]: I1010 16:04:35.927683 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.085552 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-utilities\") pod \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.085605 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv8rj\" (UniqueName: \"kubernetes.io/projected/934209ba-9aa9-4ec2-8572-37c7e924ecfe-kube-api-access-cv8rj\") pod \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.085714 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-catalog-content\") pod \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\" (UID: \"934209ba-9aa9-4ec2-8572-37c7e924ecfe\") " Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.086644 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-utilities" (OuterVolumeSpecName: "utilities") pod "934209ba-9aa9-4ec2-8572-37c7e924ecfe" (UID: "934209ba-9aa9-4ec2-8572-37c7e924ecfe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.091229 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/934209ba-9aa9-4ec2-8572-37c7e924ecfe-kube-api-access-cv8rj" (OuterVolumeSpecName: "kube-api-access-cv8rj") pod "934209ba-9aa9-4ec2-8572-37c7e924ecfe" (UID: "934209ba-9aa9-4ec2-8572-37c7e924ecfe"). InnerVolumeSpecName "kube-api-access-cv8rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.105422 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "934209ba-9aa9-4ec2-8572-37c7e924ecfe" (UID: "934209ba-9aa9-4ec2-8572-37c7e924ecfe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.187448 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.187488 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv8rj\" (UniqueName: \"kubernetes.io/projected/934209ba-9aa9-4ec2-8572-37c7e924ecfe-kube-api-access-cv8rj\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.187506 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/934209ba-9aa9-4ec2-8572-37c7e924ecfe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.518960 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" event={"ID":"436a3bba-9371-4bf0-938d-8655a4862e22","Type":"ContainerStarted","Data":"081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b"} Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.519600 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.522542 4788 generic.go:334] "Generic (PLEG): container finished" podID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerID="ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8" exitCode=0 Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.522586 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ffbxz" event={"ID":"934209ba-9aa9-4ec2-8572-37c7e924ecfe","Type":"ContainerDied","Data":"ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8"} Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.522615 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ffbxz" event={"ID":"934209ba-9aa9-4ec2-8572-37c7e924ecfe","Type":"ContainerDied","Data":"f9b8b25d1bfff1c4319d58b0dacf334c9eef44b45e057e7f735cb40d9a889569"} Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.522636 4788 scope.go:117] "RemoveContainer" containerID="ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.522653 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ffbxz" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.543200 4788 scope.go:117] "RemoveContainer" containerID="f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.545350 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" podStartSLOduration=2.545327214 podStartE2EDuration="2.545327214s" podCreationTimestamp="2025-10-10 16:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:04:36.539657803 +0000 UTC m=+4778.989373351" watchObservedRunningTime="2025-10-10 16:04:36.545327214 +0000 UTC m=+4778.995042762" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.559694 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ffbxz"] Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.564689 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ffbxz"] Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.712940 4788 scope.go:117] "RemoveContainer" containerID="a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.753442 4788 scope.go:117] "RemoveContainer" containerID="ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8" Oct 10 16:04:36 crc kubenswrapper[4788]: E1010 16:04:36.753959 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8\": container with ID starting with ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8 not found: ID does not exist" containerID="ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.753997 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8"} err="failed to get container status \"ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8\": rpc error: code = NotFound desc = could not find container \"ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8\": container with ID starting with ee70726574fda06d742789f1ab1297305f3c087807668d6da64558aafd36fff8 not found: ID does not exist" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.754020 4788 scope.go:117] "RemoveContainer" containerID="f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb" Oct 10 16:04:36 crc kubenswrapper[4788]: E1010 16:04:36.754539 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb\": container with ID starting with f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb not found: ID does not exist" containerID="f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.754601 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb"} err="failed to get container status \"f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb\": rpc error: code = NotFound desc = could not find container \"f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb\": container with ID starting with f43a123a301ca74491ce0338c1050887827d30dcac27d5346ee39072664602fb not found: ID does not exist" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.754643 4788 scope.go:117] "RemoveContainer" containerID="a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595" Oct 10 16:04:36 crc kubenswrapper[4788]: E1010 16:04:36.755239 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595\": container with ID starting with a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595 not found: ID does not exist" containerID="a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595" Oct 10 16:04:36 crc kubenswrapper[4788]: I1010 16:04:36.755279 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595"} err="failed to get container status \"a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595\": rpc error: code = NotFound desc = could not find container \"a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595\": container with ID starting with a8b875af6d3215ad200e3ce7938019b229895900eb82ab865d66dba02a02a595 not found: ID does not exist" Oct 10 16:04:37 crc kubenswrapper[4788]: I1010 16:04:37.266940 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerName="rabbitmq" containerID="cri-o://545de9b4d0601f9de022f68921e332ecc4e75a0b4a62bbefe804e3087734469f" gracePeriod=604798 Oct 10 16:04:37 crc kubenswrapper[4788]: I1010 16:04:37.810277 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerName="rabbitmq" containerID="cri-o://daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6" gracePeriod=604799 Oct 10 16:04:38 crc kubenswrapper[4788]: I1010 16:04:38.247122 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" path="/var/lib/kubelet/pods/934209ba-9aa9-4ec2-8572-37c7e924ecfe/volumes" Oct 10 16:04:39 crc kubenswrapper[4788]: I1010 16:04:39.214758 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.242:5672: connect: connection refused" Oct 10 16:04:39 crc kubenswrapper[4788]: I1010 16:04:39.574530 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.243:5672: connect: connection refused" Oct 10 16:04:43 crc kubenswrapper[4788]: I1010 16:04:43.582781 4788 generic.go:334] "Generic (PLEG): container finished" podID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerID="545de9b4d0601f9de022f68921e332ecc4e75a0b4a62bbefe804e3087734469f" exitCode=0 Oct 10 16:04:43 crc kubenswrapper[4788]: I1010 16:04:43.582978 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba","Type":"ContainerDied","Data":"545de9b4d0601f9de022f68921e332ecc4e75a0b4a62bbefe804e3087734469f"} Oct 10 16:04:43 crc kubenswrapper[4788]: I1010 16:04:43.926327 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.029954 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-erlang-cookie-secret\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.030026 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-plugins\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.030056 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-confd\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.030116 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tqgm\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-kube-api-access-5tqgm\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.030615 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.031198 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-erlang-cookie\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.031222 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-server-conf\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.031349 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.031369 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-pod-info\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.031390 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-plugins-conf\") pod \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\" (UID: \"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.031734 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.032076 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.032505 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.043651 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.043700 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-kube-api-access-5tqgm" (OuterVolumeSpecName: "kube-api-access-5tqgm") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "kube-api-access-5tqgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.054960 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-pod-info" (OuterVolumeSpecName: "pod-info") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.056903 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4" (OuterVolumeSpecName: "persistence") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.067670 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-server-conf" (OuterVolumeSpecName: "server-conf") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.127702 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" (UID: "ad2c3d8e-d28b-4c18-81f2-ed4460be64ba"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.138248 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.138292 4788 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-server-conf\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.138336 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") on node \"crc\" " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.138351 4788 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-pod-info\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.138361 4788 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.138370 4788 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.138382 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.138392 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tqgm\" (UniqueName: \"kubernetes.io/projected/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba-kube-api-access-5tqgm\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.174557 4788 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.175284 4788 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4") on node "crc" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.240022 4788 reconciler_common.go:293] "Volume detached for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.360592 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.544180 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-server-conf\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.544267 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-plugins-conf\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.545076 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.545113 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-confd\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.545218 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgpfk\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-kube-api-access-qgpfk\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.545995 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-plugins\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.546090 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075856c-3082-4eee-99fc-d283a5ee1a93-erlang-cookie-secret\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.546368 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.546563 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-erlang-cookie\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.546683 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075856c-3082-4eee-99fc-d283a5ee1a93-pod-info\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.546869 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"3075856c-3082-4eee-99fc-d283a5ee1a93\" (UID: \"3075856c-3082-4eee-99fc-d283a5ee1a93\") " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.547358 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.547445 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.547677 4788 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.548513 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-kube-api-access-qgpfk" (OuterVolumeSpecName: "kube-api-access-qgpfk") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "kube-api-access-qgpfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.552328 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3075856c-3082-4eee-99fc-d283a5ee1a93-pod-info" (OuterVolumeSpecName: "pod-info") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.553666 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3075856c-3082-4eee-99fc-d283a5ee1a93-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.557538 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e" (OuterVolumeSpecName: "persistence") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.563401 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-server-conf" (OuterVolumeSpecName: "server-conf") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.591669 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ad2c3d8e-d28b-4c18-81f2-ed4460be64ba","Type":"ContainerDied","Data":"fcae49447a8f002f1eabfea45c0bb6db2b1baf3fd85f0c3e2b9bba2554c3fa3d"} Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.591732 4788 scope.go:117] "RemoveContainer" containerID="545de9b4d0601f9de022f68921e332ecc4e75a0b4a62bbefe804e3087734469f" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.591894 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.596062 4788 generic.go:334] "Generic (PLEG): container finished" podID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerID="daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6" exitCode=0 Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.596095 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075856c-3082-4eee-99fc-d283a5ee1a93","Type":"ContainerDied","Data":"daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6"} Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.596121 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075856c-3082-4eee-99fc-d283a5ee1a93","Type":"ContainerDied","Data":"8d81c74625572dcdacee5230e950a6a48c9996de6350c771d688198a655161aa"} Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.596189 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.622022 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3075856c-3082-4eee-99fc-d283a5ee1a93" (UID: "3075856c-3082-4eee-99fc-d283a5ee1a93"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.649233 4788 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075856c-3082-4eee-99fc-d283a5ee1a93-server-conf\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.649277 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.649296 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgpfk\" (UniqueName: \"kubernetes.io/projected/3075856c-3082-4eee-99fc-d283a5ee1a93-kube-api-access-qgpfk\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.649312 4788 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075856c-3082-4eee-99fc-d283a5ee1a93-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.649329 4788 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075856c-3082-4eee-99fc-d283a5ee1a93-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.649341 4788 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075856c-3082-4eee-99fc-d283a5ee1a93-pod-info\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.649388 4788 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") on node \"crc\" " Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.672697 4788 scope.go:117] "RemoveContainer" containerID="ae5520da34fbbf08a7e447550b21b3b1b7be9875137e54a637e2e8a18d14a209" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.679054 4788 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.679500 4788 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e") on node "crc" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.683252 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.691503 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714014 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.714545 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerName="rabbitmq" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714571 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerName="rabbitmq" Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.714583 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerName="setup-container" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714593 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerName="setup-container" Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.714614 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerName="extract-utilities" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714622 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerName="extract-utilities" Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.714652 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerName="extract-content" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714660 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerName="extract-content" Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.714674 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerName="rabbitmq" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714682 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerName="rabbitmq" Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.714697 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerName="setup-container" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714706 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerName="setup-container" Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.714723 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerName="registry-server" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714731 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerName="registry-server" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714908 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" containerName="rabbitmq" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714937 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3075856c-3082-4eee-99fc-d283a5ee1a93" containerName="rabbitmq" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.714954 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="934209ba-9aa9-4ec2-8572-37c7e924ecfe" containerName="registry-server" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.716082 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.717532 4788 scope.go:117] "RemoveContainer" containerID="daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.721951 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.722209 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.722313 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-jp2xm" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.723789 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.723931 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.740301 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.740711 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.742621 4788 scope.go:117] "RemoveContainer" containerID="6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.750522 4788 reconciler_common.go:293] "Volume detached for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.769359 4788 scope.go:117] "RemoveContainer" containerID="daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6" Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.770677 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6\": container with ID starting with daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6 not found: ID does not exist" containerID="daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.770728 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6"} err="failed to get container status \"daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6\": rpc error: code = NotFound desc = could not find container \"daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6\": container with ID starting with daabe4a803dfbef49001493b5f5f92af8c3da4ec5b4c0727d94dc03af48e70f6 not found: ID does not exist" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.770759 4788 scope.go:117] "RemoveContainer" containerID="6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b" Oct 10 16:04:44 crc kubenswrapper[4788]: E1010 16:04:44.771616 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b\": container with ID starting with 6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b not found: ID does not exist" containerID="6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.771700 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b"} err="failed to get container status \"6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b\": rpc error: code = NotFound desc = could not find container \"6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b\": container with ID starting with 6bd1b10b4723a488cb1138c42d2d94d588e92f8ae9ff81c0ae35076b9bb5da8b not found: ID does not exist" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.807361 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-4kbfp"] Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.807782 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" podUID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" containerName="dnsmasq-dns" containerID="cri-o://2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead" gracePeriod=10 Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.851805 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nxh4\" (UniqueName: \"kubernetes.io/projected/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-kube-api-access-4nxh4\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.851891 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.851922 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.851945 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.851965 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.851989 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.852017 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.852040 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.852451 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.939003 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.945607 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.955973 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nxh4\" (UniqueName: \"kubernetes.io/projected/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-kube-api-access-4nxh4\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.956065 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.956093 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.956118 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.956157 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.956178 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.956209 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.956228 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.956271 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.960336 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.960802 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.961299 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.963054 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.963846 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.963889 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/499672cb3e8c066ee515fd533ccb139a382220336e9a03a35c83880afd837990/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.969026 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.970477 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.971390 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.972446 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.973532 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.974784 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.975380 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.975680 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.976249 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qcsmp" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.976556 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 10 16:04:44 crc kubenswrapper[4788]: I1010 16:04:44.987885 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nxh4\" (UniqueName: \"kubernetes.io/projected/cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec-kube-api-access-4nxh4\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.002352 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.005490 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6d8f2e34-9aec-418f-b491-b27f58ad55e4\") pod \"rabbitmq-server-0\" (UID: \"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec\") " pod="openstack/rabbitmq-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.046503 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159283 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlq2g\" (UniqueName: \"kubernetes.io/projected/f0339776-eb9a-41e2-aea6-7b16469ea1a9-kube-api-access-nlq2g\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159341 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159388 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0339776-eb9a-41e2-aea6-7b16469ea1a9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159422 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0339776-eb9a-41e2-aea6-7b16469ea1a9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159469 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0339776-eb9a-41e2-aea6-7b16469ea1a9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159502 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159522 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159562 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0339776-eb9a-41e2-aea6-7b16469ea1a9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.159590 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.210884 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261630 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlq2g\" (UniqueName: \"kubernetes.io/projected/f0339776-eb9a-41e2-aea6-7b16469ea1a9-kube-api-access-nlq2g\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261686 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261721 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0339776-eb9a-41e2-aea6-7b16469ea1a9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261751 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0339776-eb9a-41e2-aea6-7b16469ea1a9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261788 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0339776-eb9a-41e2-aea6-7b16469ea1a9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261827 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261855 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261904 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0339776-eb9a-41e2-aea6-7b16469ea1a9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.261930 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.262522 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.263300 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.266194 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0339776-eb9a-41e2-aea6-7b16469ea1a9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.266627 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0339776-eb9a-41e2-aea6-7b16469ea1a9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.268542 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.268585 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f86df4606710da878b6ae934207ef32850fa01ed5a0881365f46077ad27e410e/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.269098 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0339776-eb9a-41e2-aea6-7b16469ea1a9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.269893 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0339776-eb9a-41e2-aea6-7b16469ea1a9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.275486 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0339776-eb9a-41e2-aea6-7b16469ea1a9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.282794 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlq2g\" (UniqueName: \"kubernetes.io/projected/f0339776-eb9a-41e2-aea6-7b16469ea1a9-kube-api-access-nlq2g\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.305182 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e2cfcc73-40a3-447a-a39f-bd0a827c700e\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0339776-eb9a-41e2-aea6-7b16469ea1a9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.362687 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-dns-svc\") pod \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.362796 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-config\") pod \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.362850 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpf6p\" (UniqueName: \"kubernetes.io/projected/7f13bc87-6c3a-4620-9a08-d7c9774e7951-kube-api-access-gpf6p\") pod \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\" (UID: \"7f13bc87-6c3a-4620-9a08-d7c9774e7951\") " Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.367231 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f13bc87-6c3a-4620-9a08-d7c9774e7951-kube-api-access-gpf6p" (OuterVolumeSpecName: "kube-api-access-gpf6p") pod "7f13bc87-6c3a-4620-9a08-d7c9774e7951" (UID: "7f13bc87-6c3a-4620-9a08-d7c9774e7951"). InnerVolumeSpecName "kube-api-access-gpf6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.398569 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-config" (OuterVolumeSpecName: "config") pod "7f13bc87-6c3a-4620-9a08-d7c9774e7951" (UID: "7f13bc87-6c3a-4620-9a08-d7c9774e7951"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.400414 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f13bc87-6c3a-4620-9a08-d7c9774e7951" (UID: "7f13bc87-6c3a-4620-9a08-d7c9774e7951"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.464721 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.464771 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f13bc87-6c3a-4620-9a08-d7c9774e7951-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.464787 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpf6p\" (UniqueName: \"kubernetes.io/projected/7f13bc87-6c3a-4620-9a08-d7c9774e7951-kube-api-access-gpf6p\") on node \"crc\" DevicePath \"\"" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.551475 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.605601 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.605692 4788 generic.go:334] "Generic (PLEG): container finished" podID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" containerID="2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead" exitCode=0 Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.605768 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" event={"ID":"7f13bc87-6c3a-4620-9a08-d7c9774e7951","Type":"ContainerDied","Data":"2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead"} Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.605780 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.605805 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-4kbfp" event={"ID":"7f13bc87-6c3a-4620-9a08-d7c9774e7951","Type":"ContainerDied","Data":"523b2a05e9cd7a2621c98b826756a69e174ccc685e27e33650bf0223ddb0002b"} Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.605833 4788 scope.go:117] "RemoveContainer" containerID="2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.617366 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec","Type":"ContainerStarted","Data":"eea295f19ceafeb414eee20b1c4fff197922eda28e30416b2c692b3daf9f18fe"} Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.691264 4788 scope.go:117] "RemoveContainer" containerID="812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee" Oct 10 16:04:45 crc kubenswrapper[4788]: E1010 16:04:45.809057 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f13bc87_6c3a_4620_9a08_d7c9774e7951.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f13bc87_6c3a_4620_9a08_d7c9774e7951.slice/crio-523b2a05e9cd7a2621c98b826756a69e174ccc685e27e33650bf0223ddb0002b\": RecentStats: unable to find data in memory cache]" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.851650 4788 scope.go:117] "RemoveContainer" containerID="2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead" Oct 10 16:04:45 crc kubenswrapper[4788]: E1010 16:04:45.856599 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead\": container with ID starting with 2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead not found: ID does not exist" containerID="2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.856669 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead"} err="failed to get container status \"2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead\": rpc error: code = NotFound desc = could not find container \"2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead\": container with ID starting with 2fcd36b5fa1890cfccbd05bd06fc707c57a4633f67c5c1ffc09b43bd424c3ead not found: ID does not exist" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.856700 4788 scope.go:117] "RemoveContainer" containerID="812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee" Oct 10 16:04:45 crc kubenswrapper[4788]: E1010 16:04:45.857333 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee\": container with ID starting with 812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee not found: ID does not exist" containerID="812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.857374 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee"} err="failed to get container status \"812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee\": rpc error: code = NotFound desc = could not find container \"812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee\": container with ID starting with 812d1e564ef68cb252c7dd6ac213d4a0fd8a840215a463ad641c21240502c7ee not found: ID does not exist" Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.857763 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-4kbfp"] Oct 10 16:04:45 crc kubenswrapper[4788]: I1010 16:04:45.863037 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-4kbfp"] Oct 10 16:04:46 crc kubenswrapper[4788]: I1010 16:04:46.147278 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 10 16:04:46 crc kubenswrapper[4788]: W1010 16:04:46.152104 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0339776_eb9a_41e2_aea6_7b16469ea1a9.slice/crio-5faf616e7ccefcf99e15be52ae16278048b4483f0005a06b2c227c485fa1fc6f WatchSource:0}: Error finding container 5faf616e7ccefcf99e15be52ae16278048b4483f0005a06b2c227c485fa1fc6f: Status 404 returned error can't find the container with id 5faf616e7ccefcf99e15be52ae16278048b4483f0005a06b2c227c485fa1fc6f Oct 10 16:04:46 crc kubenswrapper[4788]: I1010 16:04:46.243100 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3075856c-3082-4eee-99fc-d283a5ee1a93" path="/var/lib/kubelet/pods/3075856c-3082-4eee-99fc-d283a5ee1a93/volumes" Oct 10 16:04:46 crc kubenswrapper[4788]: I1010 16:04:46.244085 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" path="/var/lib/kubelet/pods/7f13bc87-6c3a-4620-9a08-d7c9774e7951/volumes" Oct 10 16:04:46 crc kubenswrapper[4788]: I1010 16:04:46.245542 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad2c3d8e-d28b-4c18-81f2-ed4460be64ba" path="/var/lib/kubelet/pods/ad2c3d8e-d28b-4c18-81f2-ed4460be64ba/volumes" Oct 10 16:04:46 crc kubenswrapper[4788]: I1010 16:04:46.638829 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0339776-eb9a-41e2-aea6-7b16469ea1a9","Type":"ContainerStarted","Data":"5faf616e7ccefcf99e15be52ae16278048b4483f0005a06b2c227c485fa1fc6f"} Oct 10 16:04:47 crc kubenswrapper[4788]: I1010 16:04:47.647658 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec","Type":"ContainerStarted","Data":"070e56b0eca202c4b6ad31de0d216607dc9cfaafc09cb1182c83ad243e620beb"} Oct 10 16:04:48 crc kubenswrapper[4788]: I1010 16:04:48.665566 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0339776-eb9a-41e2-aea6-7b16469ea1a9","Type":"ContainerStarted","Data":"ecb66b5b3a0ff14b858a46089856f5fc49bf9e76680a692d6ad03396f9fff594"} Oct 10 16:05:20 crc kubenswrapper[4788]: I1010 16:05:20.951804 4788 generic.go:334] "Generic (PLEG): container finished" podID="f0339776-eb9a-41e2-aea6-7b16469ea1a9" containerID="ecb66b5b3a0ff14b858a46089856f5fc49bf9e76680a692d6ad03396f9fff594" exitCode=0 Oct 10 16:05:20 crc kubenswrapper[4788]: I1010 16:05:20.951862 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0339776-eb9a-41e2-aea6-7b16469ea1a9","Type":"ContainerDied","Data":"ecb66b5b3a0ff14b858a46089856f5fc49bf9e76680a692d6ad03396f9fff594"} Oct 10 16:05:20 crc kubenswrapper[4788]: I1010 16:05:20.955447 4788 generic.go:334] "Generic (PLEG): container finished" podID="cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec" containerID="070e56b0eca202c4b6ad31de0d216607dc9cfaafc09cb1182c83ad243e620beb" exitCode=0 Oct 10 16:05:20 crc kubenswrapper[4788]: I1010 16:05:20.955496 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec","Type":"ContainerDied","Data":"070e56b0eca202c4b6ad31de0d216607dc9cfaafc09cb1182c83ad243e620beb"} Oct 10 16:05:21 crc kubenswrapper[4788]: I1010 16:05:21.963078 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0339776-eb9a-41e2-aea6-7b16469ea1a9","Type":"ContainerStarted","Data":"1f1df421065f66143a3fce9d7c0243ad038788ffa8c28507c9027dd4a481e445"} Oct 10 16:05:21 crc kubenswrapper[4788]: I1010 16:05:21.963721 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:05:21 crc kubenswrapper[4788]: I1010 16:05:21.965527 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec","Type":"ContainerStarted","Data":"4353ffc6f8476fd6bce823ea067c8def5cdd9270b4c57a6c5fefa6826bbe4d5c"} Oct 10 16:05:21 crc kubenswrapper[4788]: I1010 16:05:21.965802 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 10 16:05:21 crc kubenswrapper[4788]: I1010 16:05:21.993424 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.993403787 podStartE2EDuration="37.993403787s" podCreationTimestamp="2025-10-10 16:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:05:21.983400522 +0000 UTC m=+4824.433116070" watchObservedRunningTime="2025-10-10 16:05:21.993403787 +0000 UTC m=+4824.443119345" Oct 10 16:05:22 crc kubenswrapper[4788]: I1010 16:05:22.010806 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.010785759 podStartE2EDuration="38.010785759s" podCreationTimestamp="2025-10-10 16:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:05:22.008254072 +0000 UTC m=+4824.457969630" watchObservedRunningTime="2025-10-10 16:05:22.010785759 +0000 UTC m=+4824.460501307" Oct 10 16:05:35 crc kubenswrapper[4788]: I1010 16:05:35.050494 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 10 16:05:35 crc kubenswrapper[4788]: I1010 16:05:35.609788 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 10 16:05:42 crc kubenswrapper[4788]: I1010 16:05:42.955206 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 10 16:05:42 crc kubenswrapper[4788]: E1010 16:05:42.956279 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" containerName="init" Oct 10 16:05:42 crc kubenswrapper[4788]: I1010 16:05:42.956293 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" containerName="init" Oct 10 16:05:42 crc kubenswrapper[4788]: E1010 16:05:42.956332 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" containerName="dnsmasq-dns" Oct 10 16:05:42 crc kubenswrapper[4788]: I1010 16:05:42.956338 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" containerName="dnsmasq-dns" Oct 10 16:05:42 crc kubenswrapper[4788]: I1010 16:05:42.956681 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f13bc87-6c3a-4620-9a08-d7c9774e7951" containerName="dnsmasq-dns" Oct 10 16:05:42 crc kubenswrapper[4788]: I1010 16:05:42.957309 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 10 16:05:42 crc kubenswrapper[4788]: I1010 16:05:42.969486 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rvhk8" Oct 10 16:05:42 crc kubenswrapper[4788]: I1010 16:05:42.973239 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 10 16:05:43 crc kubenswrapper[4788]: I1010 16:05:43.053010 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nkcj\" (UniqueName: \"kubernetes.io/projected/7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7-kube-api-access-6nkcj\") pod \"mariadb-client-1-default\" (UID: \"7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7\") " pod="openstack/mariadb-client-1-default" Oct 10 16:05:43 crc kubenswrapper[4788]: I1010 16:05:43.154458 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nkcj\" (UniqueName: \"kubernetes.io/projected/7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7-kube-api-access-6nkcj\") pod \"mariadb-client-1-default\" (UID: \"7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7\") " pod="openstack/mariadb-client-1-default" Oct 10 16:05:43 crc kubenswrapper[4788]: I1010 16:05:43.174289 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nkcj\" (UniqueName: \"kubernetes.io/projected/7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7-kube-api-access-6nkcj\") pod \"mariadb-client-1-default\" (UID: \"7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7\") " pod="openstack/mariadb-client-1-default" Oct 10 16:05:43 crc kubenswrapper[4788]: I1010 16:05:43.287125 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 10 16:05:43 crc kubenswrapper[4788]: I1010 16:05:43.791894 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 10 16:05:44 crc kubenswrapper[4788]: I1010 16:05:44.199368 4788 generic.go:334] "Generic (PLEG): container finished" podID="7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7" containerID="03191736e94a84c761d81e80f55089678532cfe6a888f55d30a9b7081eca4003" exitCode=0 Oct 10 16:05:44 crc kubenswrapper[4788]: I1010 16:05:44.199427 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7","Type":"ContainerDied","Data":"03191736e94a84c761d81e80f55089678532cfe6a888f55d30a9b7081eca4003"} Oct 10 16:05:44 crc kubenswrapper[4788]: I1010 16:05:44.199760 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7","Type":"ContainerStarted","Data":"c9d7f335af46354e0c797974082596cd7122ca1f01dec10c435587422d2e5c44"} Oct 10 16:05:45 crc kubenswrapper[4788]: I1010 16:05:45.585153 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 10 16:05:45 crc kubenswrapper[4788]: I1010 16:05:45.612845 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7/mariadb-client-1-default/0.log" Oct 10 16:05:45 crc kubenswrapper[4788]: I1010 16:05:45.638423 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 10 16:05:45 crc kubenswrapper[4788]: I1010 16:05:45.642661 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 10 16:05:45 crc kubenswrapper[4788]: I1010 16:05:45.694782 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nkcj\" (UniqueName: \"kubernetes.io/projected/7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7-kube-api-access-6nkcj\") pod \"7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7\" (UID: \"7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7\") " Oct 10 16:05:45 crc kubenswrapper[4788]: I1010 16:05:45.704455 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7-kube-api-access-6nkcj" (OuterVolumeSpecName: "kube-api-access-6nkcj") pod "7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7" (UID: "7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7"). InnerVolumeSpecName "kube-api-access-6nkcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:05:45 crc kubenswrapper[4788]: I1010 16:05:45.796885 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nkcj\" (UniqueName: \"kubernetes.io/projected/7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7-kube-api-access-6nkcj\") on node \"crc\" DevicePath \"\"" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.158707 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 10 16:05:46 crc kubenswrapper[4788]: E1010 16:05:46.159204 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7" containerName="mariadb-client-1-default" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.159229 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7" containerName="mariadb-client-1-default" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.159469 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7" containerName="mariadb-client-1-default" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.160108 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.169610 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.219092 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9d7f335af46354e0c797974082596cd7122ca1f01dec10c435587422d2e5c44" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.219175 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.242795 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7" path="/var/lib/kubelet/pods/7daf3d4d-7cf8-4c30-ad2d-0dd4534156f7/volumes" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.305422 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lqkr\" (UniqueName: \"kubernetes.io/projected/d8690a47-4da5-41b8-a8fd-486b6b09abf0-kube-api-access-4lqkr\") pod \"mariadb-client-2-default\" (UID: \"d8690a47-4da5-41b8-a8fd-486b6b09abf0\") " pod="openstack/mariadb-client-2-default" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.407182 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lqkr\" (UniqueName: \"kubernetes.io/projected/d8690a47-4da5-41b8-a8fd-486b6b09abf0-kube-api-access-4lqkr\") pod \"mariadb-client-2-default\" (UID: \"d8690a47-4da5-41b8-a8fd-486b6b09abf0\") " pod="openstack/mariadb-client-2-default" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.699914 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lqkr\" (UniqueName: \"kubernetes.io/projected/d8690a47-4da5-41b8-a8fd-486b6b09abf0-kube-api-access-4lqkr\") pod \"mariadb-client-2-default\" (UID: \"d8690a47-4da5-41b8-a8fd-486b6b09abf0\") " pod="openstack/mariadb-client-2-default" Oct 10 16:05:46 crc kubenswrapper[4788]: I1010 16:05:46.773163 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 10 16:05:47 crc kubenswrapper[4788]: I1010 16:05:47.301050 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 10 16:05:48 crc kubenswrapper[4788]: I1010 16:05:48.239029 4788 generic.go:334] "Generic (PLEG): container finished" podID="d8690a47-4da5-41b8-a8fd-486b6b09abf0" containerID="94a0a493ef9fa1802e0fa87aa8975e417efb5b4413ddf98e18aa016ea2f2fa03" exitCode=0 Oct 10 16:05:48 crc kubenswrapper[4788]: I1010 16:05:48.243476 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"d8690a47-4da5-41b8-a8fd-486b6b09abf0","Type":"ContainerDied","Data":"94a0a493ef9fa1802e0fa87aa8975e417efb5b4413ddf98e18aa016ea2f2fa03"} Oct 10 16:05:48 crc kubenswrapper[4788]: I1010 16:05:48.243723 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"d8690a47-4da5-41b8-a8fd-486b6b09abf0","Type":"ContainerStarted","Data":"4fc37016041cee8a4b35d411ba37538958020c7254542f52bf305e712a59bc5f"} Oct 10 16:05:49 crc kubenswrapper[4788]: I1010 16:05:49.571290 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 10 16:05:49 crc kubenswrapper[4788]: I1010 16:05:49.631414 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_d8690a47-4da5-41b8-a8fd-486b6b09abf0/mariadb-client-2-default/0.log" Oct 10 16:05:49 crc kubenswrapper[4788]: I1010 16:05:49.661403 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 10 16:05:49 crc kubenswrapper[4788]: I1010 16:05:49.675067 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 10 16:05:49 crc kubenswrapper[4788]: I1010 16:05:49.675593 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lqkr\" (UniqueName: \"kubernetes.io/projected/d8690a47-4da5-41b8-a8fd-486b6b09abf0-kube-api-access-4lqkr\") pod \"d8690a47-4da5-41b8-a8fd-486b6b09abf0\" (UID: \"d8690a47-4da5-41b8-a8fd-486b6b09abf0\") " Oct 10 16:05:49 crc kubenswrapper[4788]: I1010 16:05:49.694412 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8690a47-4da5-41b8-a8fd-486b6b09abf0-kube-api-access-4lqkr" (OuterVolumeSpecName: "kube-api-access-4lqkr") pod "d8690a47-4da5-41b8-a8fd-486b6b09abf0" (UID: "d8690a47-4da5-41b8-a8fd-486b6b09abf0"). InnerVolumeSpecName "kube-api-access-4lqkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:05:49 crc kubenswrapper[4788]: I1010 16:05:49.778101 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lqkr\" (UniqueName: \"kubernetes.io/projected/d8690a47-4da5-41b8-a8fd-486b6b09abf0-kube-api-access-4lqkr\") on node \"crc\" DevicePath \"\"" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.147759 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 10 16:05:50 crc kubenswrapper[4788]: E1010 16:05:50.148097 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8690a47-4da5-41b8-a8fd-486b6b09abf0" containerName="mariadb-client-2-default" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.148116 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8690a47-4da5-41b8-a8fd-486b6b09abf0" containerName="mariadb-client-2-default" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.148806 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8690a47-4da5-41b8-a8fd-486b6b09abf0" containerName="mariadb-client-2-default" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.149357 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.157453 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.245843 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8690a47-4da5-41b8-a8fd-486b6b09abf0" path="/var/lib/kubelet/pods/d8690a47-4da5-41b8-a8fd-486b6b09abf0/volumes" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.256302 4788 scope.go:117] "RemoveContainer" containerID="94a0a493ef9fa1802e0fa87aa8975e417efb5b4413ddf98e18aa016ea2f2fa03" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.256350 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.284856 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg2rg\" (UniqueName: \"kubernetes.io/projected/607fb0e6-8c1c-42d4-ba71-b803a99ec1e7-kube-api-access-qg2rg\") pod \"mariadb-client-1\" (UID: \"607fb0e6-8c1c-42d4-ba71-b803a99ec1e7\") " pod="openstack/mariadb-client-1" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.387310 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg2rg\" (UniqueName: \"kubernetes.io/projected/607fb0e6-8c1c-42d4-ba71-b803a99ec1e7-kube-api-access-qg2rg\") pod \"mariadb-client-1\" (UID: \"607fb0e6-8c1c-42d4-ba71-b803a99ec1e7\") " pod="openstack/mariadb-client-1" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.409457 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg2rg\" (UniqueName: \"kubernetes.io/projected/607fb0e6-8c1c-42d4-ba71-b803a99ec1e7-kube-api-access-qg2rg\") pod \"mariadb-client-1\" (UID: \"607fb0e6-8c1c-42d4-ba71-b803a99ec1e7\") " pod="openstack/mariadb-client-1" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.473589 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 10 16:05:50 crc kubenswrapper[4788]: I1010 16:05:50.806908 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 10 16:05:50 crc kubenswrapper[4788]: W1010 16:05:50.816709 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod607fb0e6_8c1c_42d4_ba71_b803a99ec1e7.slice/crio-8526cfb3932a1c98df64b18588e50616a4982dbb98bc6f474a21c7cd9766d2f7 WatchSource:0}: Error finding container 8526cfb3932a1c98df64b18588e50616a4982dbb98bc6f474a21c7cd9766d2f7: Status 404 returned error can't find the container with id 8526cfb3932a1c98df64b18588e50616a4982dbb98bc6f474a21c7cd9766d2f7 Oct 10 16:05:51 crc kubenswrapper[4788]: I1010 16:05:51.264029 4788 generic.go:334] "Generic (PLEG): container finished" podID="607fb0e6-8c1c-42d4-ba71-b803a99ec1e7" containerID="8a05d46dcc24ec330f690eebc82e8cc99493efc3695dee962ea7324b8b8f55e3" exitCode=0 Oct 10 16:05:51 crc kubenswrapper[4788]: I1010 16:05:51.264068 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"607fb0e6-8c1c-42d4-ba71-b803a99ec1e7","Type":"ContainerDied","Data":"8a05d46dcc24ec330f690eebc82e8cc99493efc3695dee962ea7324b8b8f55e3"} Oct 10 16:05:51 crc kubenswrapper[4788]: I1010 16:05:51.264101 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"607fb0e6-8c1c-42d4-ba71-b803a99ec1e7","Type":"ContainerStarted","Data":"8526cfb3932a1c98df64b18588e50616a4982dbb98bc6f474a21c7cd9766d2f7"} Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:52.642870 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:52.669598 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_607fb0e6-8c1c-42d4-ba71-b803a99ec1e7/mariadb-client-1/0.log" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:52.695787 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:52.701576 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:52.733629 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg2rg\" (UniqueName: \"kubernetes.io/projected/607fb0e6-8c1c-42d4-ba71-b803a99ec1e7-kube-api-access-qg2rg\") pod \"607fb0e6-8c1c-42d4-ba71-b803a99ec1e7\" (UID: \"607fb0e6-8c1c-42d4-ba71-b803a99ec1e7\") " Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:52.740247 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/607fb0e6-8c1c-42d4-ba71-b803a99ec1e7-kube-api-access-qg2rg" (OuterVolumeSpecName: "kube-api-access-qg2rg") pod "607fb0e6-8c1c-42d4-ba71-b803a99ec1e7" (UID: "607fb0e6-8c1c-42d4-ba71-b803a99ec1e7"). InnerVolumeSpecName "kube-api-access-qg2rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:52.835574 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg2rg\" (UniqueName: \"kubernetes.io/projected/607fb0e6-8c1c-42d4-ba71-b803a99ec1e7-kube-api-access-qg2rg\") on node \"crc\" DevicePath \"\"" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.229297 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 10 16:05:53 crc kubenswrapper[4788]: E1010 16:05:53.230007 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607fb0e6-8c1c-42d4-ba71-b803a99ec1e7" containerName="mariadb-client-1" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.230018 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="607fb0e6-8c1c-42d4-ba71-b803a99ec1e7" containerName="mariadb-client-1" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.230190 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="607fb0e6-8c1c-42d4-ba71-b803a99ec1e7" containerName="mariadb-client-1" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.230753 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.240580 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.287796 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8526cfb3932a1c98df64b18588e50616a4982dbb98bc6f474a21c7cd9766d2f7" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.287863 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.344192 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvbrt\" (UniqueName: \"kubernetes.io/projected/a9ae211f-d9c4-4440-b34f-912c729de737-kube-api-access-nvbrt\") pod \"mariadb-client-4-default\" (UID: \"a9ae211f-d9c4-4440-b34f-912c729de737\") " pod="openstack/mariadb-client-4-default" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.445917 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvbrt\" (UniqueName: \"kubernetes.io/projected/a9ae211f-d9c4-4440-b34f-912c729de737-kube-api-access-nvbrt\") pod \"mariadb-client-4-default\" (UID: \"a9ae211f-d9c4-4440-b34f-912c729de737\") " pod="openstack/mariadb-client-4-default" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.465030 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvbrt\" (UniqueName: \"kubernetes.io/projected/a9ae211f-d9c4-4440-b34f-912c729de737-kube-api-access-nvbrt\") pod \"mariadb-client-4-default\" (UID: \"a9ae211f-d9c4-4440-b34f-912c729de737\") " pod="openstack/mariadb-client-4-default" Oct 10 16:05:53 crc kubenswrapper[4788]: I1010 16:05:53.593375 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 10 16:05:54 crc kubenswrapper[4788]: I1010 16:05:54.012985 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 10 16:05:54 crc kubenswrapper[4788]: I1010 16:05:54.248885 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="607fb0e6-8c1c-42d4-ba71-b803a99ec1e7" path="/var/lib/kubelet/pods/607fb0e6-8c1c-42d4-ba71-b803a99ec1e7/volumes" Oct 10 16:05:54 crc kubenswrapper[4788]: I1010 16:05:54.300721 4788 generic.go:334] "Generic (PLEG): container finished" podID="a9ae211f-d9c4-4440-b34f-912c729de737" containerID="68fbbe7b711d4f1a988a33aaf62b8982887b65b73edad52551fcf6bfbf6599d1" exitCode=0 Oct 10 16:05:54 crc kubenswrapper[4788]: I1010 16:05:54.300777 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"a9ae211f-d9c4-4440-b34f-912c729de737","Type":"ContainerDied","Data":"68fbbe7b711d4f1a988a33aaf62b8982887b65b73edad52551fcf6bfbf6599d1"} Oct 10 16:05:54 crc kubenswrapper[4788]: I1010 16:05:54.300814 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"a9ae211f-d9c4-4440-b34f-912c729de737","Type":"ContainerStarted","Data":"6f8cb68025db36284b0128dd31ae7b973a4f3deca812a421c066be47b7317bd7"} Oct 10 16:05:55 crc kubenswrapper[4788]: I1010 16:05:55.651428 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 10 16:05:55 crc kubenswrapper[4788]: I1010 16:05:55.671780 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_a9ae211f-d9c4-4440-b34f-912c729de737/mariadb-client-4-default/0.log" Oct 10 16:05:55 crc kubenswrapper[4788]: I1010 16:05:55.701588 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 10 16:05:55 crc kubenswrapper[4788]: I1010 16:05:55.706148 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 10 16:05:55 crc kubenswrapper[4788]: I1010 16:05:55.794308 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvbrt\" (UniqueName: \"kubernetes.io/projected/a9ae211f-d9c4-4440-b34f-912c729de737-kube-api-access-nvbrt\") pod \"a9ae211f-d9c4-4440-b34f-912c729de737\" (UID: \"a9ae211f-d9c4-4440-b34f-912c729de737\") " Oct 10 16:05:55 crc kubenswrapper[4788]: I1010 16:05:55.800673 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ae211f-d9c4-4440-b34f-912c729de737-kube-api-access-nvbrt" (OuterVolumeSpecName: "kube-api-access-nvbrt") pod "a9ae211f-d9c4-4440-b34f-912c729de737" (UID: "a9ae211f-d9c4-4440-b34f-912c729de737"). InnerVolumeSpecName "kube-api-access-nvbrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:05:55 crc kubenswrapper[4788]: I1010 16:05:55.896392 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvbrt\" (UniqueName: \"kubernetes.io/projected/a9ae211f-d9c4-4440-b34f-912c729de737-kube-api-access-nvbrt\") on node \"crc\" DevicePath \"\"" Oct 10 16:05:56 crc kubenswrapper[4788]: I1010 16:05:56.261039 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ae211f-d9c4-4440-b34f-912c729de737" path="/var/lib/kubelet/pods/a9ae211f-d9c4-4440-b34f-912c729de737/volumes" Oct 10 16:05:56 crc kubenswrapper[4788]: I1010 16:05:56.325232 4788 scope.go:117] "RemoveContainer" containerID="68fbbe7b711d4f1a988a33aaf62b8982887b65b73edad52551fcf6bfbf6599d1" Oct 10 16:05:56 crc kubenswrapper[4788]: I1010 16:05:56.325302 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 10 16:05:59 crc kubenswrapper[4788]: I1010 16:05:59.406433 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:05:59 crc kubenswrapper[4788]: I1010 16:05:59.406965 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.338537 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 10 16:06:00 crc kubenswrapper[4788]: E1010 16:06:00.339074 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ae211f-d9c4-4440-b34f-912c729de737" containerName="mariadb-client-4-default" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.339112 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ae211f-d9c4-4440-b34f-912c729de737" containerName="mariadb-client-4-default" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.339526 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ae211f-d9c4-4440-b34f-912c729de737" containerName="mariadb-client-4-default" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.340555 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.344106 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rvhk8" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.359319 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.484876 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ds6l\" (UniqueName: \"kubernetes.io/projected/5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1-kube-api-access-9ds6l\") pod \"mariadb-client-5-default\" (UID: \"5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1\") " pod="openstack/mariadb-client-5-default" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.586878 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ds6l\" (UniqueName: \"kubernetes.io/projected/5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1-kube-api-access-9ds6l\") pod \"mariadb-client-5-default\" (UID: \"5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1\") " pod="openstack/mariadb-client-5-default" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.623552 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ds6l\" (UniqueName: \"kubernetes.io/projected/5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1-kube-api-access-9ds6l\") pod \"mariadb-client-5-default\" (UID: \"5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1\") " pod="openstack/mariadb-client-5-default" Oct 10 16:06:00 crc kubenswrapper[4788]: I1010 16:06:00.682173 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.107367 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.387637 4788 generic.go:334] "Generic (PLEG): container finished" podID="5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1" containerID="f4baf9d1df3290675f179a92f8a99c312666937fa9335670fa1f811d26854a81" exitCode=0 Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.387702 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1","Type":"ContainerDied","Data":"f4baf9d1df3290675f179a92f8a99c312666937fa9335670fa1f811d26854a81"} Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.389913 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1","Type":"ContainerStarted","Data":"83cdbaf8f8b704642a44130dd529c70c089050c7572b8a8ccedb2732688a8838"} Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.857715 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kqn6g"] Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.859782 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.877977 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kqn6g"] Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.914390 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-catalog-content\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.914460 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xftsz\" (UniqueName: \"kubernetes.io/projected/c1176d6f-3388-41fd-8a82-63a9f1aba94f-kube-api-access-xftsz\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:01 crc kubenswrapper[4788]: I1010 16:06:01.914740 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-utilities\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.017004 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-utilities\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.017133 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-catalog-content\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.017200 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xftsz\" (UniqueName: \"kubernetes.io/projected/c1176d6f-3388-41fd-8a82-63a9f1aba94f-kube-api-access-xftsz\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.017760 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-utilities\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.017820 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-catalog-content\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.049380 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xftsz\" (UniqueName: \"kubernetes.io/projected/c1176d6f-3388-41fd-8a82-63a9f1aba94f-kube-api-access-xftsz\") pod \"community-operators-kqn6g\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.188071 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.675410 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kqn6g"] Oct 10 16:06:02 crc kubenswrapper[4788]: W1010 16:06:02.680324 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1176d6f_3388_41fd_8a82_63a9f1aba94f.slice/crio-4340edc4bafe218105b182e43635029c2827a81daf8d0c80462a300d73d21bd7 WatchSource:0}: Error finding container 4340edc4bafe218105b182e43635029c2827a81daf8d0c80462a300d73d21bd7: Status 404 returned error can't find the container with id 4340edc4bafe218105b182e43635029c2827a81daf8d0c80462a300d73d21bd7 Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.848488 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.879249 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1/mariadb-client-5-default/0.log" Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.921184 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.930176 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.931545 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ds6l\" (UniqueName: \"kubernetes.io/projected/5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1-kube-api-access-9ds6l\") pod \"5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1\" (UID: \"5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1\") " Oct 10 16:06:02 crc kubenswrapper[4788]: I1010 16:06:02.943900 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1-kube-api-access-9ds6l" (OuterVolumeSpecName: "kube-api-access-9ds6l") pod "5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1" (UID: "5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1"). InnerVolumeSpecName "kube-api-access-9ds6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.033519 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ds6l\" (UniqueName: \"kubernetes.io/projected/5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1-kube-api-access-9ds6l\") on node \"crc\" DevicePath \"\"" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.125081 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 10 16:06:03 crc kubenswrapper[4788]: E1010 16:06:03.125517 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1" containerName="mariadb-client-5-default" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.125538 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1" containerName="mariadb-client-5-default" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.125734 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1" containerName="mariadb-client-5-default" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.126498 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.136264 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t2sk\" (UniqueName: \"kubernetes.io/projected/bd3cc313-d66c-4fbf-b918-ff1a8d22ce50-kube-api-access-8t2sk\") pod \"mariadb-client-6-default\" (UID: \"bd3cc313-d66c-4fbf-b918-ff1a8d22ce50\") " pod="openstack/mariadb-client-6-default" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.137766 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.237784 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t2sk\" (UniqueName: \"kubernetes.io/projected/bd3cc313-d66c-4fbf-b918-ff1a8d22ce50-kube-api-access-8t2sk\") pod \"mariadb-client-6-default\" (UID: \"bd3cc313-d66c-4fbf-b918-ff1a8d22ce50\") " pod="openstack/mariadb-client-6-default" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.257792 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t2sk\" (UniqueName: \"kubernetes.io/projected/bd3cc313-d66c-4fbf-b918-ff1a8d22ce50-kube-api-access-8t2sk\") pod \"mariadb-client-6-default\" (UID: \"bd3cc313-d66c-4fbf-b918-ff1a8d22ce50\") " pod="openstack/mariadb-client-6-default" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.412983 4788 generic.go:334] "Generic (PLEG): container finished" podID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerID="96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606" exitCode=0 Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.413057 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqn6g" event={"ID":"c1176d6f-3388-41fd-8a82-63a9f1aba94f","Type":"ContainerDied","Data":"96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606"} Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.413088 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqn6g" event={"ID":"c1176d6f-3388-41fd-8a82-63a9f1aba94f","Type":"ContainerStarted","Data":"4340edc4bafe218105b182e43635029c2827a81daf8d0c80462a300d73d21bd7"} Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.415365 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.416833 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83cdbaf8f8b704642a44130dd529c70c089050c7572b8a8ccedb2732688a8838" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.417028 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 10 16:06:03 crc kubenswrapper[4788]: I1010 16:06:03.480785 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 10 16:06:04 crc kubenswrapper[4788]: I1010 16:06:04.020225 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 10 16:06:04 crc kubenswrapper[4788]: I1010 16:06:04.243423 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1" path="/var/lib/kubelet/pods/5cc1f0cf-9865-4a54-8ea1-831eca3fc4e1/volumes" Oct 10 16:06:04 crc kubenswrapper[4788]: I1010 16:06:04.427748 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqn6g" event={"ID":"c1176d6f-3388-41fd-8a82-63a9f1aba94f","Type":"ContainerStarted","Data":"fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e"} Oct 10 16:06:04 crc kubenswrapper[4788]: I1010 16:06:04.429206 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"bd3cc313-d66c-4fbf-b918-ff1a8d22ce50","Type":"ContainerStarted","Data":"a86a76c6c542a6e7ecc652116fd806df0529053d42ca71f427e59cb8c6fb743f"} Oct 10 16:06:04 crc kubenswrapper[4788]: I1010 16:06:04.429262 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"bd3cc313-d66c-4fbf-b918-ff1a8d22ce50","Type":"ContainerStarted","Data":"628b1cdbda2cb539960a9d63ef093287fce33606fb51863208e59aca1f191764"} Oct 10 16:06:04 crc kubenswrapper[4788]: I1010 16:06:04.477893 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=1.477873529 podStartE2EDuration="1.477873529s" podCreationTimestamp="2025-10-10 16:06:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:06:04.474027757 +0000 UTC m=+4866.923743325" watchObservedRunningTime="2025-10-10 16:06:04.477873529 +0000 UTC m=+4866.927589077" Oct 10 16:06:05 crc kubenswrapper[4788]: I1010 16:06:05.441781 4788 generic.go:334] "Generic (PLEG): container finished" podID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerID="fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e" exitCode=0 Oct 10 16:06:05 crc kubenswrapper[4788]: I1010 16:06:05.441896 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqn6g" event={"ID":"c1176d6f-3388-41fd-8a82-63a9f1aba94f","Type":"ContainerDied","Data":"fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e"} Oct 10 16:06:05 crc kubenswrapper[4788]: I1010 16:06:05.445958 4788 generic.go:334] "Generic (PLEG): container finished" podID="bd3cc313-d66c-4fbf-b918-ff1a8d22ce50" containerID="a86a76c6c542a6e7ecc652116fd806df0529053d42ca71f427e59cb8c6fb743f" exitCode=0 Oct 10 16:06:05 crc kubenswrapper[4788]: I1010 16:06:05.445986 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"bd3cc313-d66c-4fbf-b918-ff1a8d22ce50","Type":"ContainerDied","Data":"a86a76c6c542a6e7ecc652116fd806df0529053d42ca71f427e59cb8c6fb743f"} Oct 10 16:06:06 crc kubenswrapper[4788]: I1010 16:06:06.464342 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqn6g" event={"ID":"c1176d6f-3388-41fd-8a82-63a9f1aba94f","Type":"ContainerStarted","Data":"a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c"} Oct 10 16:06:06 crc kubenswrapper[4788]: I1010 16:06:06.500417 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kqn6g" podStartSLOduration=2.932456383 podStartE2EDuration="5.500385279s" podCreationTimestamp="2025-10-10 16:06:01 +0000 UTC" firstStartedPulling="2025-10-10 16:06:03.415000188 +0000 UTC m=+4865.864715736" lastFinishedPulling="2025-10-10 16:06:05.982929044 +0000 UTC m=+4868.432644632" observedRunningTime="2025-10-10 16:06:06.493056744 +0000 UTC m=+4868.942772332" watchObservedRunningTime="2025-10-10 16:06:06.500385279 +0000 UTC m=+4868.950100857" Oct 10 16:06:06 crc kubenswrapper[4788]: I1010 16:06:06.846046 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 10 16:06:06 crc kubenswrapper[4788]: I1010 16:06:06.944592 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 10 16:06:06 crc kubenswrapper[4788]: I1010 16:06:06.949538 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.031925 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t2sk\" (UniqueName: \"kubernetes.io/projected/bd3cc313-d66c-4fbf-b918-ff1a8d22ce50-kube-api-access-8t2sk\") pod \"bd3cc313-d66c-4fbf-b918-ff1a8d22ce50\" (UID: \"bd3cc313-d66c-4fbf-b918-ff1a8d22ce50\") " Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.038398 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd3cc313-d66c-4fbf-b918-ff1a8d22ce50-kube-api-access-8t2sk" (OuterVolumeSpecName: "kube-api-access-8t2sk") pod "bd3cc313-d66c-4fbf-b918-ff1a8d22ce50" (UID: "bd3cc313-d66c-4fbf-b918-ff1a8d22ce50"). InnerVolumeSpecName "kube-api-access-8t2sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.134452 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t2sk\" (UniqueName: \"kubernetes.io/projected/bd3cc313-d66c-4fbf-b918-ff1a8d22ce50-kube-api-access-8t2sk\") on node \"crc\" DevicePath \"\"" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.141813 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 10 16:06:07 crc kubenswrapper[4788]: E1010 16:06:07.142173 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3cc313-d66c-4fbf-b918-ff1a8d22ce50" containerName="mariadb-client-6-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.142194 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3cc313-d66c-4fbf-b918-ff1a8d22ce50" containerName="mariadb-client-6-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.142368 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd3cc313-d66c-4fbf-b918-ff1a8d22ce50" containerName="mariadb-client-6-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.143169 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.152641 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.235803 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr7x6\" (UniqueName: \"kubernetes.io/projected/b9a780a0-8040-4cc6-a911-cfae9fb7fee4-kube-api-access-pr7x6\") pod \"mariadb-client-7-default\" (UID: \"b9a780a0-8040-4cc6-a911-cfae9fb7fee4\") " pod="openstack/mariadb-client-7-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.337332 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr7x6\" (UniqueName: \"kubernetes.io/projected/b9a780a0-8040-4cc6-a911-cfae9fb7fee4-kube-api-access-pr7x6\") pod \"mariadb-client-7-default\" (UID: \"b9a780a0-8040-4cc6-a911-cfae9fb7fee4\") " pod="openstack/mariadb-client-7-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.356328 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr7x6\" (UniqueName: \"kubernetes.io/projected/b9a780a0-8040-4cc6-a911-cfae9fb7fee4-kube-api-access-pr7x6\") pod \"mariadb-client-7-default\" (UID: \"b9a780a0-8040-4cc6-a911-cfae9fb7fee4\") " pod="openstack/mariadb-client-7-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.456595 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.485574 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="628b1cdbda2cb539960a9d63ef093287fce33606fb51863208e59aca1f191764" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.485738 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 10 16:06:07 crc kubenswrapper[4788]: I1010 16:06:07.997950 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 10 16:06:08 crc kubenswrapper[4788]: W1010 16:06:08.004080 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9a780a0_8040_4cc6_a911_cfae9fb7fee4.slice/crio-28a5ef6f6408a258380fd1284350a120391e2397c2770944c139e7f7bfe26da8 WatchSource:0}: Error finding container 28a5ef6f6408a258380fd1284350a120391e2397c2770944c139e7f7bfe26da8: Status 404 returned error can't find the container with id 28a5ef6f6408a258380fd1284350a120391e2397c2770944c139e7f7bfe26da8 Oct 10 16:06:08 crc kubenswrapper[4788]: I1010 16:06:08.244739 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd3cc313-d66c-4fbf-b918-ff1a8d22ce50" path="/var/lib/kubelet/pods/bd3cc313-d66c-4fbf-b918-ff1a8d22ce50/volumes" Oct 10 16:06:08 crc kubenswrapper[4788]: I1010 16:06:08.498856 4788 generic.go:334] "Generic (PLEG): container finished" podID="b9a780a0-8040-4cc6-a911-cfae9fb7fee4" containerID="ecb004c264fcb97dd6f299a060eb220f9f3d627e6c010d1a4b10abebc848b5f9" exitCode=0 Oct 10 16:06:08 crc kubenswrapper[4788]: I1010 16:06:08.498904 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"b9a780a0-8040-4cc6-a911-cfae9fb7fee4","Type":"ContainerDied","Data":"ecb004c264fcb97dd6f299a060eb220f9f3d627e6c010d1a4b10abebc848b5f9"} Oct 10 16:06:08 crc kubenswrapper[4788]: I1010 16:06:08.498961 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"b9a780a0-8040-4cc6-a911-cfae9fb7fee4","Type":"ContainerStarted","Data":"28a5ef6f6408a258380fd1284350a120391e2397c2770944c139e7f7bfe26da8"} Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.104542 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.127747 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_b9a780a0-8040-4cc6-a911-cfae9fb7fee4/mariadb-client-7-default/0.log" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.156720 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.165276 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.289801 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr7x6\" (UniqueName: \"kubernetes.io/projected/b9a780a0-8040-4cc6-a911-cfae9fb7fee4-kube-api-access-pr7x6\") pod \"b9a780a0-8040-4cc6-a911-cfae9fb7fee4\" (UID: \"b9a780a0-8040-4cc6-a911-cfae9fb7fee4\") " Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.298951 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a780a0-8040-4cc6-a911-cfae9fb7fee4-kube-api-access-pr7x6" (OuterVolumeSpecName: "kube-api-access-pr7x6") pod "b9a780a0-8040-4cc6-a911-cfae9fb7fee4" (UID: "b9a780a0-8040-4cc6-a911-cfae9fb7fee4"). InnerVolumeSpecName "kube-api-access-pr7x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.347313 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 10 16:06:10 crc kubenswrapper[4788]: E1010 16:06:10.347823 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a780a0-8040-4cc6-a911-cfae9fb7fee4" containerName="mariadb-client-7-default" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.347848 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a780a0-8040-4cc6-a911-cfae9fb7fee4" containerName="mariadb-client-7-default" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.348074 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a780a0-8040-4cc6-a911-cfae9fb7fee4" containerName="mariadb-client-7-default" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.348852 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.360045 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.391353 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr7x6\" (UniqueName: \"kubernetes.io/projected/b9a780a0-8040-4cc6-a911-cfae9fb7fee4-kube-api-access-pr7x6\") on node \"crc\" DevicePath \"\"" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.493325 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5z6x\" (UniqueName: \"kubernetes.io/projected/615cf126-69d9-48cb-872a-f2096ca90f69-kube-api-access-b5z6x\") pod \"mariadb-client-2\" (UID: \"615cf126-69d9-48cb-872a-f2096ca90f69\") " pod="openstack/mariadb-client-2" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.515909 4788 scope.go:117] "RemoveContainer" containerID="ecb004c264fcb97dd6f299a060eb220f9f3d627e6c010d1a4b10abebc848b5f9" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.516003 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.594399 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5z6x\" (UniqueName: \"kubernetes.io/projected/615cf126-69d9-48cb-872a-f2096ca90f69-kube-api-access-b5z6x\") pod \"mariadb-client-2\" (UID: \"615cf126-69d9-48cb-872a-f2096ca90f69\") " pod="openstack/mariadb-client-2" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.610530 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5z6x\" (UniqueName: \"kubernetes.io/projected/615cf126-69d9-48cb-872a-f2096ca90f69-kube-api-access-b5z6x\") pod \"mariadb-client-2\" (UID: \"615cf126-69d9-48cb-872a-f2096ca90f69\") " pod="openstack/mariadb-client-2" Oct 10 16:06:10 crc kubenswrapper[4788]: I1010 16:06:10.676131 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 10 16:06:11 crc kubenswrapper[4788]: I1010 16:06:11.223555 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 10 16:06:11 crc kubenswrapper[4788]: W1010 16:06:11.227781 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod615cf126_69d9_48cb_872a_f2096ca90f69.slice/crio-f44fd87905e867eef3df3e8222385c9322d8d2c389535f66596fe2069d522d7e WatchSource:0}: Error finding container f44fd87905e867eef3df3e8222385c9322d8d2c389535f66596fe2069d522d7e: Status 404 returned error can't find the container with id f44fd87905e867eef3df3e8222385c9322d8d2c389535f66596fe2069d522d7e Oct 10 16:06:11 crc kubenswrapper[4788]: I1010 16:06:11.543101 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"615cf126-69d9-48cb-872a-f2096ca90f69","Type":"ContainerStarted","Data":"f44fd87905e867eef3df3e8222385c9322d8d2c389535f66596fe2069d522d7e"} Oct 10 16:06:12 crc kubenswrapper[4788]: I1010 16:06:12.189122 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:12 crc kubenswrapper[4788]: I1010 16:06:12.189225 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:12 crc kubenswrapper[4788]: I1010 16:06:12.250451 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a780a0-8040-4cc6-a911-cfae9fb7fee4" path="/var/lib/kubelet/pods/b9a780a0-8040-4cc6-a911-cfae9fb7fee4/volumes" Oct 10 16:06:12 crc kubenswrapper[4788]: I1010 16:06:12.270785 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:12 crc kubenswrapper[4788]: I1010 16:06:12.558404 4788 generic.go:334] "Generic (PLEG): container finished" podID="615cf126-69d9-48cb-872a-f2096ca90f69" containerID="137cfb1a2e2482cd3c5be2f7433082725ac3e4e41a4b84147434e8c887c2fee3" exitCode=0 Oct 10 16:06:12 crc kubenswrapper[4788]: I1010 16:06:12.558469 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"615cf126-69d9-48cb-872a-f2096ca90f69","Type":"ContainerDied","Data":"137cfb1a2e2482cd3c5be2f7433082725ac3e4e41a4b84147434e8c887c2fee3"} Oct 10 16:06:12 crc kubenswrapper[4788]: I1010 16:06:12.635440 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:12 crc kubenswrapper[4788]: I1010 16:06:12.701244 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kqn6g"] Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.024234 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.051891 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_615cf126-69d9-48cb-872a-f2096ca90f69/mariadb-client-2/0.log" Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.065473 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5z6x\" (UniqueName: \"kubernetes.io/projected/615cf126-69d9-48cb-872a-f2096ca90f69-kube-api-access-b5z6x\") pod \"615cf126-69d9-48cb-872a-f2096ca90f69\" (UID: \"615cf126-69d9-48cb-872a-f2096ca90f69\") " Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.079340 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/615cf126-69d9-48cb-872a-f2096ca90f69-kube-api-access-b5z6x" (OuterVolumeSpecName: "kube-api-access-b5z6x") pod "615cf126-69d9-48cb-872a-f2096ca90f69" (UID: "615cf126-69d9-48cb-872a-f2096ca90f69"). InnerVolumeSpecName "kube-api-access-b5z6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.089246 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.096572 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.167872 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5z6x\" (UniqueName: \"kubernetes.io/projected/615cf126-69d9-48cb-872a-f2096ca90f69-kube-api-access-b5z6x\") on node \"crc\" DevicePath \"\"" Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.250826 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="615cf126-69d9-48cb-872a-f2096ca90f69" path="/var/lib/kubelet/pods/615cf126-69d9-48cb-872a-f2096ca90f69/volumes" Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.583396 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.583433 4788 scope.go:117] "RemoveContainer" containerID="137cfb1a2e2482cd3c5be2f7433082725ac3e4e41a4b84147434e8c887c2fee3" Oct 10 16:06:14 crc kubenswrapper[4788]: I1010 16:06:14.584175 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kqn6g" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerName="registry-server" containerID="cri-o://a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c" gracePeriod=2 Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.069558 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.185535 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-utilities\") pod \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.185638 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xftsz\" (UniqueName: \"kubernetes.io/projected/c1176d6f-3388-41fd-8a82-63a9f1aba94f-kube-api-access-xftsz\") pod \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.185734 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-catalog-content\") pod \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\" (UID: \"c1176d6f-3388-41fd-8a82-63a9f1aba94f\") " Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.187834 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-utilities" (OuterVolumeSpecName: "utilities") pod "c1176d6f-3388-41fd-8a82-63a9f1aba94f" (UID: "c1176d6f-3388-41fd-8a82-63a9f1aba94f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.193563 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1176d6f-3388-41fd-8a82-63a9f1aba94f-kube-api-access-xftsz" (OuterVolumeSpecName: "kube-api-access-xftsz") pod "c1176d6f-3388-41fd-8a82-63a9f1aba94f" (UID: "c1176d6f-3388-41fd-8a82-63a9f1aba94f"). InnerVolumeSpecName "kube-api-access-xftsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.247687 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1176d6f-3388-41fd-8a82-63a9f1aba94f" (UID: "c1176d6f-3388-41fd-8a82-63a9f1aba94f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.288295 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.288507 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xftsz\" (UniqueName: \"kubernetes.io/projected/c1176d6f-3388-41fd-8a82-63a9f1aba94f-kube-api-access-xftsz\") on node \"crc\" DevicePath \"\"" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.289317 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1176d6f-3388-41fd-8a82-63a9f1aba94f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.596550 4788 generic.go:334] "Generic (PLEG): container finished" podID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerID="a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c" exitCode=0 Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.596586 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqn6g" event={"ID":"c1176d6f-3388-41fd-8a82-63a9f1aba94f","Type":"ContainerDied","Data":"a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c"} Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.596607 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kqn6g" event={"ID":"c1176d6f-3388-41fd-8a82-63a9f1aba94f","Type":"ContainerDied","Data":"4340edc4bafe218105b182e43635029c2827a81daf8d0c80462a300d73d21bd7"} Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.596625 4788 scope.go:117] "RemoveContainer" containerID="a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.596730 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kqn6g" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.636776 4788 scope.go:117] "RemoveContainer" containerID="fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.648130 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kqn6g"] Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.659220 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kqn6g"] Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.659823 4788 scope.go:117] "RemoveContainer" containerID="96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.719984 4788 scope.go:117] "RemoveContainer" containerID="a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c" Oct 10 16:06:15 crc kubenswrapper[4788]: E1010 16:06:15.720839 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c\": container with ID starting with a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c not found: ID does not exist" containerID="a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.720898 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c"} err="failed to get container status \"a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c\": rpc error: code = NotFound desc = could not find container \"a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c\": container with ID starting with a4258bc521d6853ad420e6e37520e0f73231cd5bf1c5344adb990891d263050c not found: ID does not exist" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.720937 4788 scope.go:117] "RemoveContainer" containerID="fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e" Oct 10 16:06:15 crc kubenswrapper[4788]: E1010 16:06:15.721366 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e\": container with ID starting with fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e not found: ID does not exist" containerID="fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.721408 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e"} err="failed to get container status \"fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e\": rpc error: code = NotFound desc = could not find container \"fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e\": container with ID starting with fd3b3aeb9ef7b486f1049f75e6893fa9c35fd13ec343ef3f5a99ae6b4a7ef18e not found: ID does not exist" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.721445 4788 scope.go:117] "RemoveContainer" containerID="96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606" Oct 10 16:06:15 crc kubenswrapper[4788]: E1010 16:06:15.722169 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606\": container with ID starting with 96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606 not found: ID does not exist" containerID="96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606" Oct 10 16:06:15 crc kubenswrapper[4788]: I1010 16:06:15.722207 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606"} err="failed to get container status \"96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606\": rpc error: code = NotFound desc = could not find container \"96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606\": container with ID starting with 96c93055620a68d421100d06287ecdcd4babd46a2a45c328cf700f7deacbd606 not found: ID does not exist" Oct 10 16:06:16 crc kubenswrapper[4788]: I1010 16:06:16.261262 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" path="/var/lib/kubelet/pods/c1176d6f-3388-41fd-8a82-63a9f1aba94f/volumes" Oct 10 16:06:29 crc kubenswrapper[4788]: I1010 16:06:29.406528 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:06:29 crc kubenswrapper[4788]: I1010 16:06:29.407436 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:06:59 crc kubenswrapper[4788]: I1010 16:06:59.405968 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:06:59 crc kubenswrapper[4788]: I1010 16:06:59.406631 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:06:59 crc kubenswrapper[4788]: I1010 16:06:59.406693 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:06:59 crc kubenswrapper[4788]: I1010 16:06:59.407741 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:06:59 crc kubenswrapper[4788]: I1010 16:06:59.407826 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" gracePeriod=600 Oct 10 16:06:59 crc kubenswrapper[4788]: E1010 16:06:59.553481 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:07:00 crc kubenswrapper[4788]: I1010 16:07:00.063379 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" exitCode=0 Oct 10 16:07:00 crc kubenswrapper[4788]: I1010 16:07:00.063451 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e"} Oct 10 16:07:00 crc kubenswrapper[4788]: I1010 16:07:00.063545 4788 scope.go:117] "RemoveContainer" containerID="6df1945d6fe1182ef3cdc4c7a6a92c339bc56228bcd7417f29306d9a6b290161" Oct 10 16:07:00 crc kubenswrapper[4788]: I1010 16:07:00.065304 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:07:00 crc kubenswrapper[4788]: E1010 16:07:00.066491 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:07:12 crc kubenswrapper[4788]: I1010 16:07:12.041747 4788 scope.go:117] "RemoveContainer" containerID="edc563279ab190b7eabed0297e85d99da4dfa732ddbbcf2c0cd277a319ef7ba3" Oct 10 16:07:12 crc kubenswrapper[4788]: I1010 16:07:12.234794 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:07:12 crc kubenswrapper[4788]: E1010 16:07:12.235434 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:07:23 crc kubenswrapper[4788]: I1010 16:07:23.235698 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:07:23 crc kubenswrapper[4788]: E1010 16:07:23.237411 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:07:38 crc kubenswrapper[4788]: I1010 16:07:38.242186 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:07:38 crc kubenswrapper[4788]: E1010 16:07:38.243417 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:07:50 crc kubenswrapper[4788]: I1010 16:07:50.234681 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:07:50 crc kubenswrapper[4788]: E1010 16:07:50.235769 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:08:02 crc kubenswrapper[4788]: I1010 16:08:02.234376 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:08:02 crc kubenswrapper[4788]: E1010 16:08:02.235505 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:08:16 crc kubenswrapper[4788]: I1010 16:08:16.234496 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:08:16 crc kubenswrapper[4788]: E1010 16:08:16.235500 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:08:27 crc kubenswrapper[4788]: I1010 16:08:27.234790 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:08:27 crc kubenswrapper[4788]: E1010 16:08:27.236128 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:08:41 crc kubenswrapper[4788]: I1010 16:08:41.235060 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:08:41 crc kubenswrapper[4788]: E1010 16:08:41.236184 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:08:54 crc kubenswrapper[4788]: I1010 16:08:54.234850 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:08:54 crc kubenswrapper[4788]: E1010 16:08:54.236245 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:09:09 crc kubenswrapper[4788]: I1010 16:09:09.235415 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:09:09 crc kubenswrapper[4788]: E1010 16:09:09.237093 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:09:20 crc kubenswrapper[4788]: I1010 16:09:20.233834 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:09:20 crc kubenswrapper[4788]: E1010 16:09:20.234879 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:09:35 crc kubenswrapper[4788]: I1010 16:09:35.233961 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:09:35 crc kubenswrapper[4788]: E1010 16:09:35.235108 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:09:46 crc kubenswrapper[4788]: I1010 16:09:46.235631 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:09:46 crc kubenswrapper[4788]: E1010 16:09:46.236974 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:09:59 crc kubenswrapper[4788]: I1010 16:09:59.234220 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:09:59 crc kubenswrapper[4788]: E1010 16:09:59.235210 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:10:12 crc kubenswrapper[4788]: I1010 16:10:12.234745 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:10:12 crc kubenswrapper[4788]: E1010 16:10:12.236091 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:10:27 crc kubenswrapper[4788]: I1010 16:10:27.234165 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:10:27 crc kubenswrapper[4788]: E1010 16:10:27.234816 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:10:42 crc kubenswrapper[4788]: I1010 16:10:42.235534 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:10:42 crc kubenswrapper[4788]: E1010 16:10:42.237404 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:10:57 crc kubenswrapper[4788]: I1010 16:10:57.233716 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:10:57 crc kubenswrapper[4788]: E1010 16:10:57.234557 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.564941 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 10 16:11:04 crc kubenswrapper[4788]: E1010 16:11:04.565697 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerName="extract-utilities" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.565709 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerName="extract-utilities" Oct 10 16:11:04 crc kubenswrapper[4788]: E1010 16:11:04.565725 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerName="registry-server" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.565730 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerName="registry-server" Oct 10 16:11:04 crc kubenswrapper[4788]: E1010 16:11:04.565741 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="615cf126-69d9-48cb-872a-f2096ca90f69" containerName="mariadb-client-2" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.565748 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="615cf126-69d9-48cb-872a-f2096ca90f69" containerName="mariadb-client-2" Oct 10 16:11:04 crc kubenswrapper[4788]: E1010 16:11:04.565771 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerName="extract-content" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.565777 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerName="extract-content" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.565912 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="615cf126-69d9-48cb-872a-f2096ca90f69" containerName="mariadb-client-2" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.565927 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1176d6f-3388-41fd-8a82-63a9f1aba94f" containerName="registry-server" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.566371 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.575492 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rvhk8" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.576552 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.695433 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-10f888c9-d148-409f-9038-656d64d3677f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10f888c9-d148-409f-9038-656d64d3677f\") pod \"mariadb-copy-data\" (UID: \"8b0029af-55ed-45fc-b831-d115439a4591\") " pod="openstack/mariadb-copy-data" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.695512 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55frk\" (UniqueName: \"kubernetes.io/projected/8b0029af-55ed-45fc-b831-d115439a4591-kube-api-access-55frk\") pod \"mariadb-copy-data\" (UID: \"8b0029af-55ed-45fc-b831-d115439a4591\") " pod="openstack/mariadb-copy-data" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.797237 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55frk\" (UniqueName: \"kubernetes.io/projected/8b0029af-55ed-45fc-b831-d115439a4591-kube-api-access-55frk\") pod \"mariadb-copy-data\" (UID: \"8b0029af-55ed-45fc-b831-d115439a4591\") " pod="openstack/mariadb-copy-data" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.797370 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-10f888c9-d148-409f-9038-656d64d3677f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10f888c9-d148-409f-9038-656d64d3677f\") pod \"mariadb-copy-data\" (UID: \"8b0029af-55ed-45fc-b831-d115439a4591\") " pod="openstack/mariadb-copy-data" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.821257 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55frk\" (UniqueName: \"kubernetes.io/projected/8b0029af-55ed-45fc-b831-d115439a4591-kube-api-access-55frk\") pod \"mariadb-copy-data\" (UID: \"8b0029af-55ed-45fc-b831-d115439a4591\") " pod="openstack/mariadb-copy-data" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.832490 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.832840 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-10f888c9-d148-409f-9038-656d64d3677f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10f888c9-d148-409f-9038-656d64d3677f\") pod \"mariadb-copy-data\" (UID: \"8b0029af-55ed-45fc-b831-d115439a4591\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8408951588db0029aa31c6bc316473e8f4260772f34a83e2d940c2f54bd3a292/globalmount\"" pod="openstack/mariadb-copy-data" Oct 10 16:11:04 crc kubenswrapper[4788]: I1010 16:11:04.935536 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-10f888c9-d148-409f-9038-656d64d3677f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-10f888c9-d148-409f-9038-656d64d3677f\") pod \"mariadb-copy-data\" (UID: \"8b0029af-55ed-45fc-b831-d115439a4591\") " pod="openstack/mariadb-copy-data" Oct 10 16:11:05 crc kubenswrapper[4788]: I1010 16:11:05.191327 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 10 16:11:05 crc kubenswrapper[4788]: I1010 16:11:05.733181 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 10 16:11:06 crc kubenswrapper[4788]: I1010 16:11:06.500416 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"8b0029af-55ed-45fc-b831-d115439a4591","Type":"ContainerStarted","Data":"c9af2220d3bee117c92ffc07661f973e9b7231c437aa05c383fe96a8daa6eaaa"} Oct 10 16:11:06 crc kubenswrapper[4788]: I1010 16:11:06.500460 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"8b0029af-55ed-45fc-b831-d115439a4591","Type":"ContainerStarted","Data":"7bd45cbfc3fb5ac51bd37fa7436369668aecd535ba7d9d250f4f83e979ca730a"} Oct 10 16:11:06 crc kubenswrapper[4788]: I1010 16:11:06.515583 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.515568195 podStartE2EDuration="3.515568195s" podCreationTimestamp="2025-10-10 16:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:06.512728102 +0000 UTC m=+5168.962443650" watchObservedRunningTime="2025-10-10 16:11:06.515568195 +0000 UTC m=+5168.965283743" Oct 10 16:11:08 crc kubenswrapper[4788]: I1010 16:11:08.537318 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:08 crc kubenswrapper[4788]: I1010 16:11:08.538718 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 10 16:11:08 crc kubenswrapper[4788]: I1010 16:11:08.553348 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:08 crc kubenswrapper[4788]: I1010 16:11:08.666481 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t25ts\" (UniqueName: \"kubernetes.io/projected/ca542871-ab5e-4dd0-a398-40d558a8fe78-kube-api-access-t25ts\") pod \"mariadb-client\" (UID: \"ca542871-ab5e-4dd0-a398-40d558a8fe78\") " pod="openstack/mariadb-client" Oct 10 16:11:08 crc kubenswrapper[4788]: I1010 16:11:08.768751 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t25ts\" (UniqueName: \"kubernetes.io/projected/ca542871-ab5e-4dd0-a398-40d558a8fe78-kube-api-access-t25ts\") pod \"mariadb-client\" (UID: \"ca542871-ab5e-4dd0-a398-40d558a8fe78\") " pod="openstack/mariadb-client" Oct 10 16:11:08 crc kubenswrapper[4788]: I1010 16:11:08.790116 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t25ts\" (UniqueName: \"kubernetes.io/projected/ca542871-ab5e-4dd0-a398-40d558a8fe78-kube-api-access-t25ts\") pod \"mariadb-client\" (UID: \"ca542871-ab5e-4dd0-a398-40d558a8fe78\") " pod="openstack/mariadb-client" Oct 10 16:11:08 crc kubenswrapper[4788]: I1010 16:11:08.861872 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 10 16:11:09 crc kubenswrapper[4788]: I1010 16:11:09.292346 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:09 crc kubenswrapper[4788]: W1010 16:11:09.297691 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca542871_ab5e_4dd0_a398_40d558a8fe78.slice/crio-8d9ba0325c14352935317f8a5b0fae07bb666bdba0ea62f5869614c4e95124c4 WatchSource:0}: Error finding container 8d9ba0325c14352935317f8a5b0fae07bb666bdba0ea62f5869614c4e95124c4: Status 404 returned error can't find the container with id 8d9ba0325c14352935317f8a5b0fae07bb666bdba0ea62f5869614c4e95124c4 Oct 10 16:11:09 crc kubenswrapper[4788]: I1010 16:11:09.535240 4788 generic.go:334] "Generic (PLEG): container finished" podID="ca542871-ab5e-4dd0-a398-40d558a8fe78" containerID="1113508caceb0d522af8209c17bf099d5e088853b3299f5a3c49843e65d5d1e9" exitCode=0 Oct 10 16:11:09 crc kubenswrapper[4788]: I1010 16:11:09.535299 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"ca542871-ab5e-4dd0-a398-40d558a8fe78","Type":"ContainerDied","Data":"1113508caceb0d522af8209c17bf099d5e088853b3299f5a3c49843e65d5d1e9"} Oct 10 16:11:09 crc kubenswrapper[4788]: I1010 16:11:09.535683 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"ca542871-ab5e-4dd0-a398-40d558a8fe78","Type":"ContainerStarted","Data":"8d9ba0325c14352935317f8a5b0fae07bb666bdba0ea62f5869614c4e95124c4"} Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.798631 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.819981 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_ca542871-ab5e-4dd0-a398-40d558a8fe78/mariadb-client/0.log" Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.850153 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.857586 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.898436 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t25ts\" (UniqueName: \"kubernetes.io/projected/ca542871-ab5e-4dd0-a398-40d558a8fe78-kube-api-access-t25ts\") pod \"ca542871-ab5e-4dd0-a398-40d558a8fe78\" (UID: \"ca542871-ab5e-4dd0-a398-40d558a8fe78\") " Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.902808 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca542871-ab5e-4dd0-a398-40d558a8fe78-kube-api-access-t25ts" (OuterVolumeSpecName: "kube-api-access-t25ts") pod "ca542871-ab5e-4dd0-a398-40d558a8fe78" (UID: "ca542871-ab5e-4dd0-a398-40d558a8fe78"). InnerVolumeSpecName "kube-api-access-t25ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.990530 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:10 crc kubenswrapper[4788]: E1010 16:11:10.990825 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca542871-ab5e-4dd0-a398-40d558a8fe78" containerName="mariadb-client" Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.990842 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca542871-ab5e-4dd0-a398-40d558a8fe78" containerName="mariadb-client" Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.990990 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca542871-ab5e-4dd0-a398-40d558a8fe78" containerName="mariadb-client" Oct 10 16:11:10 crc kubenswrapper[4788]: I1010 16:11:10.991464 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.000775 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t25ts\" (UniqueName: \"kubernetes.io/projected/ca542871-ab5e-4dd0-a398-40d558a8fe78-kube-api-access-t25ts\") on node \"crc\" DevicePath \"\"" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.005071 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.102186 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctmsf\" (UniqueName: \"kubernetes.io/projected/71121393-55ac-4f78-8077-b1a175ed3a07-kube-api-access-ctmsf\") pod \"mariadb-client\" (UID: \"71121393-55ac-4f78-8077-b1a175ed3a07\") " pod="openstack/mariadb-client" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.207793 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctmsf\" (UniqueName: \"kubernetes.io/projected/71121393-55ac-4f78-8077-b1a175ed3a07-kube-api-access-ctmsf\") pod \"mariadb-client\" (UID: \"71121393-55ac-4f78-8077-b1a175ed3a07\") " pod="openstack/mariadb-client" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.224059 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctmsf\" (UniqueName: \"kubernetes.io/projected/71121393-55ac-4f78-8077-b1a175ed3a07-kube-api-access-ctmsf\") pod \"mariadb-client\" (UID: \"71121393-55ac-4f78-8077-b1a175ed3a07\") " pod="openstack/mariadb-client" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.235033 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:11:11 crc kubenswrapper[4788]: E1010 16:11:11.236647 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.316050 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.552503 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d9ba0325c14352935317f8a5b0fae07bb666bdba0ea62f5869614c4e95124c4" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.552644 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.571987 4788 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="ca542871-ab5e-4dd0-a398-40d558a8fe78" podUID="71121393-55ac-4f78-8077-b1a175ed3a07" Oct 10 16:11:11 crc kubenswrapper[4788]: W1010 16:11:11.764304 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71121393_55ac_4f78_8077_b1a175ed3a07.slice/crio-53c92b56fca90120ce19bccb2287ab24b957913add804d24c2a853180967f45a WatchSource:0}: Error finding container 53c92b56fca90120ce19bccb2287ab24b957913add804d24c2a853180967f45a: Status 404 returned error can't find the container with id 53c92b56fca90120ce19bccb2287ab24b957913add804d24c2a853180967f45a Oct 10 16:11:11 crc kubenswrapper[4788]: I1010 16:11:11.765105 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:12 crc kubenswrapper[4788]: I1010 16:11:12.245370 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca542871-ab5e-4dd0-a398-40d558a8fe78" path="/var/lib/kubelet/pods/ca542871-ab5e-4dd0-a398-40d558a8fe78/volumes" Oct 10 16:11:12 crc kubenswrapper[4788]: I1010 16:11:12.561924 4788 generic.go:334] "Generic (PLEG): container finished" podID="71121393-55ac-4f78-8077-b1a175ed3a07" containerID="dbdd0d048754a7a280d80c58bb771cdaa9ae8f1690d1a51ac645d33ed825ee43" exitCode=0 Oct 10 16:11:12 crc kubenswrapper[4788]: I1010 16:11:12.561974 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"71121393-55ac-4f78-8077-b1a175ed3a07","Type":"ContainerDied","Data":"dbdd0d048754a7a280d80c58bb771cdaa9ae8f1690d1a51ac645d33ed825ee43"} Oct 10 16:11:12 crc kubenswrapper[4788]: I1010 16:11:12.562005 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"71121393-55ac-4f78-8077-b1a175ed3a07","Type":"ContainerStarted","Data":"53c92b56fca90120ce19bccb2287ab24b957913add804d24c2a853180967f45a"} Oct 10 16:11:13 crc kubenswrapper[4788]: I1010 16:11:13.905422 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 10 16:11:13 crc kubenswrapper[4788]: I1010 16:11:13.924410 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_71121393-55ac-4f78-8077-b1a175ed3a07/mariadb-client/0.log" Oct 10 16:11:13 crc kubenswrapper[4788]: I1010 16:11:13.950797 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:13 crc kubenswrapper[4788]: I1010 16:11:13.955894 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 10 16:11:13 crc kubenswrapper[4788]: I1010 16:11:13.956320 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctmsf\" (UniqueName: \"kubernetes.io/projected/71121393-55ac-4f78-8077-b1a175ed3a07-kube-api-access-ctmsf\") pod \"71121393-55ac-4f78-8077-b1a175ed3a07\" (UID: \"71121393-55ac-4f78-8077-b1a175ed3a07\") " Oct 10 16:11:13 crc kubenswrapper[4788]: I1010 16:11:13.962198 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71121393-55ac-4f78-8077-b1a175ed3a07-kube-api-access-ctmsf" (OuterVolumeSpecName: "kube-api-access-ctmsf") pod "71121393-55ac-4f78-8077-b1a175ed3a07" (UID: "71121393-55ac-4f78-8077-b1a175ed3a07"). InnerVolumeSpecName "kube-api-access-ctmsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:11:14 crc kubenswrapper[4788]: I1010 16:11:14.058385 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctmsf\" (UniqueName: \"kubernetes.io/projected/71121393-55ac-4f78-8077-b1a175ed3a07-kube-api-access-ctmsf\") on node \"crc\" DevicePath \"\"" Oct 10 16:11:14 crc kubenswrapper[4788]: I1010 16:11:14.241224 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71121393-55ac-4f78-8077-b1a175ed3a07" path="/var/lib/kubelet/pods/71121393-55ac-4f78-8077-b1a175ed3a07/volumes" Oct 10 16:11:14 crc kubenswrapper[4788]: E1010 16:11:14.321019 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71121393_55ac_4f78_8077_b1a175ed3a07.slice\": RecentStats: unable to find data in memory cache]" Oct 10 16:11:14 crc kubenswrapper[4788]: I1010 16:11:14.577130 4788 scope.go:117] "RemoveContainer" containerID="dbdd0d048754a7a280d80c58bb771cdaa9ae8f1690d1a51ac645d33ed825ee43" Oct 10 16:11:14 crc kubenswrapper[4788]: I1010 16:11:14.577169 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 10 16:11:24 crc kubenswrapper[4788]: I1010 16:11:24.233657 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:11:24 crc kubenswrapper[4788]: E1010 16:11:24.234748 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:11:36 crc kubenswrapper[4788]: I1010 16:11:36.240968 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:11:36 crc kubenswrapper[4788]: E1010 16:11:36.242860 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:11:43 crc kubenswrapper[4788]: E1010 16:11:43.471669 4788 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.129:45950->38.102.83.129:41727: write tcp 38.102.83.129:45950->38.102.83.129:41727: write: broken pipe Oct 10 16:11:44 crc kubenswrapper[4788]: E1010 16:11:44.247391 4788 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.129:46012->38.102.83.129:41727: write tcp 38.102.83.129:46012->38.102.83.129:41727: write: broken pipe Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.337363 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 16:11:46 crc kubenswrapper[4788]: E1010 16:11:46.338013 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71121393-55ac-4f78-8077-b1a175ed3a07" containerName="mariadb-client" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.338032 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="71121393-55ac-4f78-8077-b1a175ed3a07" containerName="mariadb-client" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.338223 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="71121393-55ac-4f78-8077-b1a175ed3a07" containerName="mariadb-client" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.339964 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.347096 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.347476 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-dskf2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.349905 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.351942 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.369292 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.380007 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.392938 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.395229 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.430126 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.434669 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.516062 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.518342 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.520319 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-mv6vr" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.520737 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.521980 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538431 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/892d027b-e0f3-420b-a4db-09926ccbf288-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538545 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e313987-6d14-4ea1-8ba6-616a31065713-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538615 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-12a4ed55-b7b9-4501-8d0e-379db0a01415\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12a4ed55-b7b9-4501-8d0e-379db0a01415\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538638 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/892d027b-e0f3-420b-a4db-09926ccbf288-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538673 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-config\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538695 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e313987-6d14-4ea1-8ba6-616a31065713-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538757 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538803 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7ac29198-991f-460b-b49a-dd5279ffb036\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7ac29198-991f-460b-b49a-dd5279ffb036\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538854 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e313987-6d14-4ea1-8ba6-616a31065713-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538876 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538908 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538932 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbr7t\" (UniqueName: \"kubernetes.io/projected/892d027b-e0f3-420b-a4db-09926ccbf288-kube-api-access-gbr7t\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.538978 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e313987-6d14-4ea1-8ba6-616a31065713-config\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.539008 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/892d027b-e0f3-420b-a4db-09926ccbf288-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.539051 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-67c54e2b-b954-4c3b-b3dd-af11c80e505b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-67c54e2b-b954-4c3b-b3dd-af11c80e505b\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.539079 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/892d027b-e0f3-420b-a4db-09926ccbf288-config\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.539108 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87wjm\" (UniqueName: \"kubernetes.io/projected/5e313987-6d14-4ea1-8ba6-616a31065713-kube-api-access-87wjm\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.539204 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrjbw\" (UniqueName: \"kubernetes.io/projected/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-kube-api-access-mrjbw\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.541861 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.552228 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.554019 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.562496 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.564220 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.572382 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.592486 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.641306 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e313987-6d14-4ea1-8ba6-616a31065713-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.641395 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-12a4ed55-b7b9-4501-8d0e-379db0a01415\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12a4ed55-b7b9-4501-8d0e-379db0a01415\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.641422 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/892d027b-e0f3-420b-a4db-09926ccbf288-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.641462 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.641486 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-config\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.641502 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e313987-6d14-4ea1-8ba6-616a31065713-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.641526 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rjxg\" (UniqueName: \"kubernetes.io/projected/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-kube-api-access-2rjxg\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.641566 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.642632 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e313987-6d14-4ea1-8ba6-616a31065713-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.642719 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7ac29198-991f-460b-b49a-dd5279ffb036\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7ac29198-991f-460b-b49a-dd5279ffb036\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.642811 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e313987-6d14-4ea1-8ba6-616a31065713-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.642837 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.642868 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.642901 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbr7t\" (UniqueName: \"kubernetes.io/projected/892d027b-e0f3-420b-a4db-09926ccbf288-kube-api-access-gbr7t\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.642948 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e313987-6d14-4ea1-8ba6-616a31065713-config\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.642988 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643013 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/892d027b-e0f3-420b-a4db-09926ccbf288-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643040 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643062 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643080 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-67c54e2b-b954-4c3b-b3dd-af11c80e505b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-67c54e2b-b954-4c3b-b3dd-af11c80e505b\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643128 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/892d027b-e0f3-420b-a4db-09926ccbf288-config\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643189 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87wjm\" (UniqueName: \"kubernetes.io/projected/5e313987-6d14-4ea1-8ba6-616a31065713-kube-api-access-87wjm\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643235 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-config\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643248 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrjbw\" (UniqueName: \"kubernetes.io/projected/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-kube-api-access-mrjbw\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643616 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c1ec5a66-3b06-464a-a178-09011920ed3a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c1ec5a66-3b06-464a-a178-09011920ed3a\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643684 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/892d027b-e0f3-420b-a4db-09926ccbf288-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.643711 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-config\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.644020 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e313987-6d14-4ea1-8ba6-616a31065713-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.644046 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/892d027b-e0f3-420b-a4db-09926ccbf288-config\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.644339 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e313987-6d14-4ea1-8ba6-616a31065713-config\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.644622 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/892d027b-e0f3-420b-a4db-09926ccbf288-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.644684 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.644757 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/892d027b-e0f3-420b-a4db-09926ccbf288-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.647043 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.647059 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.647087 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7ac29198-991f-460b-b49a-dd5279ffb036\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7ac29198-991f-460b-b49a-dd5279ffb036\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bb8050a7aabd4385f97500d0d358db749f9fbf666dcc8042e2d9e5ad4a11e3df/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.647097 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-67c54e2b-b954-4c3b-b3dd-af11c80e505b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-67c54e2b-b954-4c3b-b3dd-af11c80e505b\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/16b7c387db63e31f8db611e393b134a906fc54714b142cf43a111dc564b2c6b8/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.649024 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.649065 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-12a4ed55-b7b9-4501-8d0e-379db0a01415\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12a4ed55-b7b9-4501-8d0e-379db0a01415\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9ed2820d0414a35bc3fa34cfd544d7fe789126db7a6d0c1acbc39061f9d6e707/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.650318 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.650656 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/892d027b-e0f3-420b-a4db-09926ccbf288-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.660504 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e313987-6d14-4ea1-8ba6-616a31065713-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.667958 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrjbw\" (UniqueName: \"kubernetes.io/projected/98ebdaa9-c26a-406d-b56f-68c4a9cf8da3-kube-api-access-mrjbw\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.669592 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbr7t\" (UniqueName: \"kubernetes.io/projected/892d027b-e0f3-420b-a4db-09926ccbf288-kube-api-access-gbr7t\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.677096 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87wjm\" (UniqueName: \"kubernetes.io/projected/5e313987-6d14-4ea1-8ba6-616a31065713-kube-api-access-87wjm\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.690403 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-67c54e2b-b954-4c3b-b3dd-af11c80e505b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-67c54e2b-b954-4c3b-b3dd-af11c80e505b\") pod \"ovsdbserver-sb-0\" (UID: \"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3\") " pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.692262 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7ac29198-991f-460b-b49a-dd5279ffb036\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7ac29198-991f-460b-b49a-dd5279ffb036\") pod \"ovsdbserver-sb-2\" (UID: \"5e313987-6d14-4ea1-8ba6-616a31065713\") " pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.699534 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-12a4ed55-b7b9-4501-8d0e-379db0a01415\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12a4ed55-b7b9-4501-8d0e-379db0a01415\") pod \"ovsdbserver-sb-1\" (UID: \"892d027b-e0f3-420b-a4db-09926ccbf288\") " pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.710776 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.739038 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.744886 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4006913e-1f61-4abb-857f-af4822963e7e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4006913e-1f61-4abb-857f-af4822963e7e\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.744960 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.744998 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745027 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk76v\" (UniqueName: \"kubernetes.io/projected/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-kube-api-access-qk76v\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745077 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p67ng\" (UniqueName: \"kubernetes.io/projected/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-kube-api-access-p67ng\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745123 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745172 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745229 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-config\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745267 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7187a86d-fd46-4adc-a66f-d8e6e6df2987\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7187a86d-fd46-4adc-a66f-d8e6e6df2987\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745296 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-config\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745356 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c1ec5a66-3b06-464a-a178-09011920ed3a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c1ec5a66-3b06-464a-a178-09011920ed3a\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745403 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-config\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745431 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745478 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745501 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745527 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745555 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rjxg\" (UniqueName: \"kubernetes.io/projected/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-kube-api-access-2rjxg\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.745587 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.747714 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-config\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.748646 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.748874 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.753076 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.753130 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c1ec5a66-3b06-464a-a178-09011920ed3a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c1ec5a66-3b06-464a-a178-09011920ed3a\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/844881def81445efb10b9996738373f9dae5b3bfa1bae893be330542c07caa48/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.753768 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.771185 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rjxg\" (UniqueName: \"kubernetes.io/projected/6d7c4590-0091-4af7-b1cf-98c2afbf1cbb-kube-api-access-2rjxg\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.787723 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c1ec5a66-3b06-464a-a178-09011920ed3a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c1ec5a66-3b06-464a-a178-09011920ed3a\") pod \"ovsdbserver-nb-0\" (UID: \"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb\") " pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.846257 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847560 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847611 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847643 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk76v\" (UniqueName: \"kubernetes.io/projected/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-kube-api-access-qk76v\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847705 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p67ng\" (UniqueName: \"kubernetes.io/projected/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-kube-api-access-p67ng\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847777 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-config\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847811 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7187a86d-fd46-4adc-a66f-d8e6e6df2987\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7187a86d-fd46-4adc-a66f-d8e6e6df2987\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847839 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-config\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847886 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847928 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.847955 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.848001 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.848035 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4006913e-1f61-4abb-857f-af4822963e7e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4006913e-1f61-4abb-857f-af4822963e7e\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.848969 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.849388 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.849416 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.850233 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-config\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.850248 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.850275 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-config\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.854998 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.855055 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4006913e-1f61-4abb-857f-af4822963e7e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4006913e-1f61-4abb-857f-af4822963e7e\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/83330dc5953e5763d8465da8b58712db7c574d10ecd5fcbfa93ba41a3f72dbbb/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.855005 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.855173 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7187a86d-fd46-4adc-a66f-d8e6e6df2987\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7187a86d-fd46-4adc-a66f-d8e6e6df2987\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3c4da0662fa26bc9e50e5bca4d91f94565a8da0b25572f58d3188de29f2fd18a/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.858540 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.860044 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.870867 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p67ng\" (UniqueName: \"kubernetes.io/projected/237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2-kube-api-access-p67ng\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.872188 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk76v\" (UniqueName: \"kubernetes.io/projected/6e5012dd-51b6-4db7-8371-b71f2cc46ec3-kube-api-access-qk76v\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.903106 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7187a86d-fd46-4adc-a66f-d8e6e6df2987\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7187a86d-fd46-4adc-a66f-d8e6e6df2987\") pod \"ovsdbserver-nb-2\" (UID: \"6e5012dd-51b6-4db7-8371-b71f2cc46ec3\") " pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.904884 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4006913e-1f61-4abb-857f-af4822963e7e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4006913e-1f61-4abb-857f-af4822963e7e\") pod \"ovsdbserver-nb-1\" (UID: \"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2\") " pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:46 crc kubenswrapper[4788]: I1010 16:11:46.993240 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.187929 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.195287 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.199289 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.297225 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 10 16:11:47 crc kubenswrapper[4788]: W1010 16:11:47.305003 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod892d027b_e0f3_420b_a4db_09926ccbf288.slice/crio-968ec7167a1eab9f01895b4899e9ca23896a1dd52d601a0c6c4f50b4a3420fb1 WatchSource:0}: Error finding container 968ec7167a1eab9f01895b4899e9ca23896a1dd52d601a0c6c4f50b4a3420fb1: Status 404 returned error can't find the container with id 968ec7167a1eab9f01895b4899e9ca23896a1dd52d601a0c6c4f50b4a3420fb1 Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.405302 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 10 16:11:47 crc kubenswrapper[4788]: W1010 16:11:47.435432 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98ebdaa9_c26a_406d_b56f_68c4a9cf8da3.slice/crio-f9de10e33b42e044c21eb3c57eb5e74689dfdc1d255158998329fbe18ef0ebb3 WatchSource:0}: Error finding container f9de10e33b42e044c21eb3c57eb5e74689dfdc1d255158998329fbe18ef0ebb3: Status 404 returned error can't find the container with id f9de10e33b42e044c21eb3c57eb5e74689dfdc1d255158998329fbe18ef0ebb3 Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.797576 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.916444 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 10 16:11:47 crc kubenswrapper[4788]: W1010 16:11:47.918184 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod237d4343_a4d1_42c4_bc8b_1e49ebd7f6a2.slice/crio-0cb0ed486d99173f507b57d3c71a2f2363d8d64f710f7927bc78ffd34f2853bf WatchSource:0}: Error finding container 0cb0ed486d99173f507b57d3c71a2f2363d8d64f710f7927bc78ffd34f2853bf: Status 404 returned error can't find the container with id 0cb0ed486d99173f507b57d3c71a2f2363d8d64f710f7927bc78ffd34f2853bf Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.919958 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"6e5012dd-51b6-4db7-8371-b71f2cc46ec3","Type":"ContainerStarted","Data":"f4a7c14628e604048563a035256d6f47fe699cbdc7b3c3d2c830a797029d4f51"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.923690 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3","Type":"ContainerStarted","Data":"b2eb82c52ae0d70b0879fbeee4d5a88bde1e3c0ea5816fc5bdf4a13bb44413ae"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.923752 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3","Type":"ContainerStarted","Data":"3f0e3f1ddd42be68a5be82fd4809b40c4860b55804df4880cae7ad72c92c68c0"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.923766 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"98ebdaa9-c26a-406d-b56f-68c4a9cf8da3","Type":"ContainerStarted","Data":"f9de10e33b42e044c21eb3c57eb5e74689dfdc1d255158998329fbe18ef0ebb3"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.926089 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb","Type":"ContainerStarted","Data":"7b2c779d7e89e05bf99223a56e6f60aa2280e02519fbbe51e98f58fc8fd4ea98"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.926225 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb","Type":"ContainerStarted","Data":"a38a16dd4724e31b145565f16978ab3c6cd434eccf693f0af1bdd3c07c741b77"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.926236 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6d7c4590-0091-4af7-b1cf-98c2afbf1cbb","Type":"ContainerStarted","Data":"6157e6c0961774e1d55e424b7c9e881c67d96b197295012bb9a4b38f7fda875d"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.931894 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"892d027b-e0f3-420b-a4db-09926ccbf288","Type":"ContainerStarted","Data":"55bc985d9c7526ddf06275d6ff968817f78f8122c79ca2981080d503353928c8"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.931928 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"892d027b-e0f3-420b-a4db-09926ccbf288","Type":"ContainerStarted","Data":"6abd28793ac11fc94a5342d9eadaeddffa2e06f3fcdeab92369658c533cbf680"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.931940 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"892d027b-e0f3-420b-a4db-09926ccbf288","Type":"ContainerStarted","Data":"968ec7167a1eab9f01895b4899e9ca23896a1dd52d601a0c6c4f50b4a3420fb1"} Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.956169 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=2.956134576 podStartE2EDuration="2.956134576s" podCreationTimestamp="2025-10-10 16:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:47.944383563 +0000 UTC m=+5210.394099111" watchObservedRunningTime="2025-10-10 16:11:47.956134576 +0000 UTC m=+5210.405850124" Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.964480 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=2.96446508 podStartE2EDuration="2.96446508s" podCreationTimestamp="2025-10-10 16:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:47.963566956 +0000 UTC m=+5210.413282504" watchObservedRunningTime="2025-10-10 16:11:47.96446508 +0000 UTC m=+5210.414180628" Oct 10 16:11:47 crc kubenswrapper[4788]: I1010 16:11:47.992457 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=2.992435648 podStartE2EDuration="2.992435648s" podCreationTimestamp="2025-10-10 16:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:47.9866949 +0000 UTC m=+5210.436410628" watchObservedRunningTime="2025-10-10 16:11:47.992435648 +0000 UTC m=+5210.442151206" Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.386779 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 10 16:11:48 crc kubenswrapper[4788]: W1010 16:11:48.391619 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e313987_6d14_4ea1_8ba6_616a31065713.slice/crio-edb56c291b199dfae8caf1199ab2bd9b17080680a0400ac4e03cfd7569d24fb6 WatchSource:0}: Error finding container edb56c291b199dfae8caf1199ab2bd9b17080680a0400ac4e03cfd7569d24fb6: Status 404 returned error can't find the container with id edb56c291b199dfae8caf1199ab2bd9b17080680a0400ac4e03cfd7569d24fb6 Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.953216 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"6e5012dd-51b6-4db7-8371-b71f2cc46ec3","Type":"ContainerStarted","Data":"82270e962cf43b8839978c1c3118ecc7af521e83270e39f53e33bf9121e28b6d"} Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.953854 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"6e5012dd-51b6-4db7-8371-b71f2cc46ec3","Type":"ContainerStarted","Data":"49e2455289bc5501714c959b6338a3a2a507a8bcfbe274b6411de00c7d1e8553"} Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.958404 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2","Type":"ContainerStarted","Data":"487d62352a03cfa76be4afaa3d77dac42b3bb3cc82961ec27d1a6ea554da019f"} Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.958467 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2","Type":"ContainerStarted","Data":"a4366e6499e4007ed7a59cd4763082f05ef44d33ff7d351e05356c6ede41a6ef"} Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.958491 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2","Type":"ContainerStarted","Data":"0cb0ed486d99173f507b57d3c71a2f2363d8d64f710f7927bc78ffd34f2853bf"} Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.962068 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"5e313987-6d14-4ea1-8ba6-616a31065713","Type":"ContainerStarted","Data":"6f6a435380c95d4b7a573f4c7fd7a8f7fd8b50cca59fed0afc8ab61baa3337e1"} Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.962234 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"5e313987-6d14-4ea1-8ba6-616a31065713","Type":"ContainerStarted","Data":"2e7b09bcecc55341b263e18a1fa91a1c661510298e2f73b727f369d2d933b0f7"} Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.962273 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"5e313987-6d14-4ea1-8ba6-616a31065713","Type":"ContainerStarted","Data":"edb56c291b199dfae8caf1199ab2bd9b17080680a0400ac4e03cfd7569d24fb6"} Oct 10 16:11:48 crc kubenswrapper[4788]: I1010 16:11:48.995592 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.995547888 podStartE2EDuration="3.995547888s" podCreationTimestamp="2025-10-10 16:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:48.985478849 +0000 UTC m=+5211.435194427" watchObservedRunningTime="2025-10-10 16:11:48.995547888 +0000 UTC m=+5211.445263466" Oct 10 16:11:49 crc kubenswrapper[4788]: I1010 16:11:49.022769 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.022714335 podStartE2EDuration="4.022714335s" podCreationTimestamp="2025-10-10 16:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:49.014894705 +0000 UTC m=+5211.464610283" watchObservedRunningTime="2025-10-10 16:11:49.022714335 +0000 UTC m=+5211.472429933" Oct 10 16:11:49 crc kubenswrapper[4788]: I1010 16:11:49.711465 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:49 crc kubenswrapper[4788]: I1010 16:11:49.739841 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:49 crc kubenswrapper[4788]: I1010 16:11:49.846357 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:49 crc kubenswrapper[4788]: I1010 16:11:49.993894 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:50 crc kubenswrapper[4788]: I1010 16:11:50.189323 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:50 crc kubenswrapper[4788]: I1010 16:11:50.196773 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:50 crc kubenswrapper[4788]: I1010 16:11:50.234321 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:11:50 crc kubenswrapper[4788]: E1010 16:11:50.234736 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:11:50 crc kubenswrapper[4788]: I1010 16:11:50.253058 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:50 crc kubenswrapper[4788]: I1010 16:11:50.290331 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=5.290296939 podStartE2EDuration="5.290296939s" podCreationTimestamp="2025-10-10 16:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:49.043284224 +0000 UTC m=+5211.492999772" watchObservedRunningTime="2025-10-10 16:11:50.290296939 +0000 UTC m=+5212.740012517" Oct 10 16:11:50 crc kubenswrapper[4788]: I1010 16:11:50.979251 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:51 crc kubenswrapper[4788]: I1010 16:11:51.711372 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:51 crc kubenswrapper[4788]: I1010 16:11:51.739382 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:51 crc kubenswrapper[4788]: I1010 16:11:51.846464 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:51 crc kubenswrapper[4788]: I1010 16:11:51.994335 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.195988 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.255494 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.589586 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ccc69b885-9rx4l"] Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.591707 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.594311 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.598109 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccc69b885-9rx4l"] Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.664312 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-ovsdbserver-nb\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.664463 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-dns-svc\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.664542 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqw7b\" (UniqueName: \"kubernetes.io/projected/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-kube-api-access-nqw7b\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.664589 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-config\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.756031 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.766339 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-ovsdbserver-nb\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.766454 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-dns-svc\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.766530 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqw7b\" (UniqueName: \"kubernetes.io/projected/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-kube-api-access-nqw7b\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.766606 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-config\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.767792 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-config\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.767818 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-dns-svc\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.767842 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-ovsdbserver-nb\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.786522 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.805229 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqw7b\" (UniqueName: \"kubernetes.io/projected/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-kube-api-access-nqw7b\") pod \"dnsmasq-dns-7ccc69b885-9rx4l\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.810325 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.890335 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.936935 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:52 crc kubenswrapper[4788]: I1010 16:11:52.970930 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.088101 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccc69b885-9rx4l"] Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.091070 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.118567 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-ccf97d949-s6wcz"] Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.124845 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.127240 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.130306 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.132611 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-ccf97d949-s6wcz"] Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.202333 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.317072 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b7p5\" (UniqueName: \"kubernetes.io/projected/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-kube-api-access-8b7p5\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.317155 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-config\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.317226 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-sb\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.317257 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-nb\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.317304 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-dns-svc\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.318850 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.419604 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-sb\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.419672 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-nb\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.419725 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-dns-svc\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.419810 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b7p5\" (UniqueName: \"kubernetes.io/projected/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-kube-api-access-8b7p5\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.419838 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-config\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.421065 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-sb\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.421631 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-nb\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.422210 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-dns-svc\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.423537 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-config\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.466369 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b7p5\" (UniqueName: \"kubernetes.io/projected/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-kube-api-access-8b7p5\") pod \"dnsmasq-dns-ccf97d949-s6wcz\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.468339 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.498002 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.564575 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccc69b885-9rx4l"] Oct 10 16:11:53 crc kubenswrapper[4788]: I1010 16:11:53.946958 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-ccf97d949-s6wcz"] Oct 10 16:11:53 crc kubenswrapper[4788]: W1010 16:11:53.972195 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae76828a_d10f_49d2_94ae_39c6cd67ea0c.slice/crio-12650234e6cc244db8fcd26091d15f1836df97ecf91fa61feb7f845e8041035b WatchSource:0}: Error finding container 12650234e6cc244db8fcd26091d15f1836df97ecf91fa61feb7f845e8041035b: Status 404 returned error can't find the container with id 12650234e6cc244db8fcd26091d15f1836df97ecf91fa61feb7f845e8041035b Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.029380 4788 generic.go:334] "Generic (PLEG): container finished" podID="f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" containerID="19230b892bfdb2846049e31593532f17365d8b922282cd4503e5f7379680d096" exitCode=0 Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.029681 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" event={"ID":"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a","Type":"ContainerDied","Data":"19230b892bfdb2846049e31593532f17365d8b922282cd4503e5f7379680d096"} Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.029794 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" event={"ID":"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a","Type":"ContainerStarted","Data":"eda1970f3a002ef95b1ed72fc97b6778a741b263cb5083c5c2b09f9cab414dd4"} Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.031895 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" event={"ID":"ae76828a-d10f-49d2-94ae-39c6cd67ea0c","Type":"ContainerStarted","Data":"12650234e6cc244db8fcd26091d15f1836df97ecf91fa61feb7f845e8041035b"} Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.291626 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.336583 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqw7b\" (UniqueName: \"kubernetes.io/projected/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-kube-api-access-nqw7b\") pod \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.336630 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-ovsdbserver-nb\") pod \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.336711 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-config\") pod \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.336763 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-dns-svc\") pod \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\" (UID: \"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a\") " Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.345068 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-kube-api-access-nqw7b" (OuterVolumeSpecName: "kube-api-access-nqw7b") pod "f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" (UID: "f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a"). InnerVolumeSpecName "kube-api-access-nqw7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.362419 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-config" (OuterVolumeSpecName: "config") pod "f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" (UID: "f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.373682 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" (UID: "f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.379237 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" (UID: "f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.438362 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.438398 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqw7b\" (UniqueName: \"kubernetes.io/projected/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-kube-api-access-nqw7b\") on node \"crc\" DevicePath \"\"" Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.438411 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:11:54 crc kubenswrapper[4788]: I1010 16:11:54.438420 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:11:55 crc kubenswrapper[4788]: I1010 16:11:55.047958 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" event={"ID":"f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a","Type":"ContainerDied","Data":"eda1970f3a002ef95b1ed72fc97b6778a741b263cb5083c5c2b09f9cab414dd4"} Oct 10 16:11:55 crc kubenswrapper[4788]: I1010 16:11:55.048021 4788 scope.go:117] "RemoveContainer" containerID="19230b892bfdb2846049e31593532f17365d8b922282cd4503e5f7379680d096" Oct 10 16:11:55 crc kubenswrapper[4788]: I1010 16:11:55.048109 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc69b885-9rx4l" Oct 10 16:11:55 crc kubenswrapper[4788]: I1010 16:11:55.052032 4788 generic.go:334] "Generic (PLEG): container finished" podID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" containerID="75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6" exitCode=0 Oct 10 16:11:55 crc kubenswrapper[4788]: I1010 16:11:55.052115 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" event={"ID":"ae76828a-d10f-49d2-94ae-39c6cd67ea0c","Type":"ContainerDied","Data":"75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6"} Oct 10 16:11:55 crc kubenswrapper[4788]: I1010 16:11:55.243834 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccc69b885-9rx4l"] Oct 10 16:11:55 crc kubenswrapper[4788]: I1010 16:11:55.248689 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ccc69b885-9rx4l"] Oct 10 16:11:55 crc kubenswrapper[4788]: E1010 16:11:55.283820 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf586f5b5_c6f7_4d7d_81e1_ae18fbd4dc0a.slice/crio-eda1970f3a002ef95b1ed72fc97b6778a741b263cb5083c5c2b09f9cab414dd4\": RecentStats: unable to find data in memory cache]" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.060642 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" event={"ID":"ae76828a-d10f-49d2-94ae-39c6cd67ea0c","Type":"ContainerStarted","Data":"227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0"} Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.061266 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.096795 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" podStartSLOduration=3.096777626 podStartE2EDuration="3.096777626s" podCreationTimestamp="2025-10-10 16:11:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:56.094109498 +0000 UTC m=+5218.543825056" watchObservedRunningTime="2025-10-10 16:11:56.096777626 +0000 UTC m=+5218.546493174" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.247446 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" path="/var/lib/kubelet/pods/f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a/volumes" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.572425 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 10 16:11:56 crc kubenswrapper[4788]: E1010 16:11:56.572908 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" containerName="init" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.572922 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" containerName="init" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.573134 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f586f5b5-c6f7-4d7d-81e1-ae18fbd4dc0a" containerName="init" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.573845 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.579681 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.599311 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.776693 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/dfa6f880-c8b3-49a4-9397-022761316eb3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.776803 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxx5q\" (UniqueName: \"kubernetes.io/projected/dfa6f880-c8b3-49a4-9397-022761316eb3-kube-api-access-zxx5q\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.776833 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ba653a98-1b36-419a-ae1c-870a1ba76e92\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba653a98-1b36-419a-ae1c-870a1ba76e92\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.878169 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxx5q\" (UniqueName: \"kubernetes.io/projected/dfa6f880-c8b3-49a4-9397-022761316eb3-kube-api-access-zxx5q\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.878247 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ba653a98-1b36-419a-ae1c-870a1ba76e92\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba653a98-1b36-419a-ae1c-870a1ba76e92\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.878334 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/dfa6f880-c8b3-49a4-9397-022761316eb3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.896118 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/dfa6f880-c8b3-49a4-9397-022761316eb3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.898047 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.898124 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ba653a98-1b36-419a-ae1c-870a1ba76e92\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba653a98-1b36-419a-ae1c-870a1ba76e92\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/703212cf96d1c68c90eb47ddc138292020e2389234f7a028d6d2c785a0153eb9/globalmount\"" pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.910299 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxx5q\" (UniqueName: \"kubernetes.io/projected/dfa6f880-c8b3-49a4-9397-022761316eb3-kube-api-access-zxx5q\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:56 crc kubenswrapper[4788]: I1010 16:11:56.965295 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ba653a98-1b36-419a-ae1c-870a1ba76e92\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba653a98-1b36-419a-ae1c-870a1ba76e92\") pod \"ovn-copy-data\" (UID: \"dfa6f880-c8b3-49a4-9397-022761316eb3\") " pod="openstack/ovn-copy-data" Oct 10 16:11:57 crc kubenswrapper[4788]: I1010 16:11:57.197272 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 10 16:11:57 crc kubenswrapper[4788]: I1010 16:11:57.727203 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 10 16:11:58 crc kubenswrapper[4788]: I1010 16:11:58.087449 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"dfa6f880-c8b3-49a4-9397-022761316eb3","Type":"ContainerStarted","Data":"7eeda4e97f0c80beabfeccd40f00fdd8ac63a4c7b9b5e8ced2cf2f7da0913ed5"} Oct 10 16:11:58 crc kubenswrapper[4788]: I1010 16:11:58.087500 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"dfa6f880-c8b3-49a4-9397-022761316eb3","Type":"ContainerStarted","Data":"20553444f094932e920222f6ced96a6e10444abd8441863e22ea122e46a71e16"} Oct 10 16:11:58 crc kubenswrapper[4788]: I1010 16:11:58.114338 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.114315467 podStartE2EDuration="3.114315467s" podCreationTimestamp="2025-10-10 16:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:11:58.108552379 +0000 UTC m=+5220.558267987" watchObservedRunningTime="2025-10-10 16:11:58.114315467 +0000 UTC m=+5220.564031005" Oct 10 16:12:01 crc kubenswrapper[4788]: I1010 16:12:01.234730 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:12:02 crc kubenswrapper[4788]: I1010 16:12:02.131261 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"f2a936fbf95226f44505aa4c234e52e71830372cdc6c6f705ee31bf2d04b44c1"} Oct 10 16:12:03 crc kubenswrapper[4788]: I1010 16:12:03.470349 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:12:03 crc kubenswrapper[4788]: I1010 16:12:03.533602 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-92hfq"] Oct 10 16:12:03 crc kubenswrapper[4788]: I1010 16:12:03.533816 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" podUID="436a3bba-9371-4bf0-938d-8655a4862e22" containerName="dnsmasq-dns" containerID="cri-o://081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b" gracePeriod=10 Oct 10 16:12:03 crc kubenswrapper[4788]: I1010 16:12:03.997031 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.056671 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 10 16:12:04 crc kubenswrapper[4788]: E1010 16:12:04.057038 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436a3bba-9371-4bf0-938d-8655a4862e22" containerName="dnsmasq-dns" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.057058 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="436a3bba-9371-4bf0-938d-8655a4862e22" containerName="dnsmasq-dns" Oct 10 16:12:04 crc kubenswrapper[4788]: E1010 16:12:04.057081 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436a3bba-9371-4bf0-938d-8655a4862e22" containerName="init" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.057089 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="436a3bba-9371-4bf0-938d-8655a4862e22" containerName="init" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.057297 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="436a3bba-9371-4bf0-938d-8655a4862e22" containerName="dnsmasq-dns" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.060735 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.064716 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.065039 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.065473 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-64x6p" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.081848 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.109699 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-dns-svc\") pod \"436a3bba-9371-4bf0-938d-8655a4862e22\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.109928 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9ssw\" (UniqueName: \"kubernetes.io/projected/436a3bba-9371-4bf0-938d-8655a4862e22-kube-api-access-j9ssw\") pod \"436a3bba-9371-4bf0-938d-8655a4862e22\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.110162 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-config\") pod \"436a3bba-9371-4bf0-938d-8655a4862e22\" (UID: \"436a3bba-9371-4bf0-938d-8655a4862e22\") " Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.130454 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/436a3bba-9371-4bf0-938d-8655a4862e22-kube-api-access-j9ssw" (OuterVolumeSpecName: "kube-api-access-j9ssw") pod "436a3bba-9371-4bf0-938d-8655a4862e22" (UID: "436a3bba-9371-4bf0-938d-8655a4862e22"). InnerVolumeSpecName "kube-api-access-j9ssw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.158418 4788 generic.go:334] "Generic (PLEG): container finished" podID="436a3bba-9371-4bf0-938d-8655a4862e22" containerID="081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b" exitCode=0 Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.158461 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" event={"ID":"436a3bba-9371-4bf0-938d-8655a4862e22","Type":"ContainerDied","Data":"081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b"} Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.158487 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" event={"ID":"436a3bba-9371-4bf0-938d-8655a4862e22","Type":"ContainerDied","Data":"bc7e81f9cc68f3615c84815044e3cf9562cd489a4ef457f8598f903d43ef7abe"} Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.158509 4788 scope.go:117] "RemoveContainer" containerID="081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.158634 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-92hfq" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.164871 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-config" (OuterVolumeSpecName: "config") pod "436a3bba-9371-4bf0-938d-8655a4862e22" (UID: "436a3bba-9371-4bf0-938d-8655a4862e22"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.167704 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "436a3bba-9371-4bf0-938d-8655a4862e22" (UID: "436a3bba-9371-4bf0-938d-8655a4862e22"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.195878 4788 scope.go:117] "RemoveContainer" containerID="cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.212877 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.212977 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.213038 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5scg\" (UniqueName: \"kubernetes.io/projected/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-kube-api-access-k5scg\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.213059 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-config\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.213083 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-scripts\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.213152 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.213164 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/436a3bba-9371-4bf0-938d-8655a4862e22-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.213176 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9ssw\" (UniqueName: \"kubernetes.io/projected/436a3bba-9371-4bf0-938d-8655a4862e22-kube-api-access-j9ssw\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.217360 4788 scope.go:117] "RemoveContainer" containerID="081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b" Oct 10 16:12:04 crc kubenswrapper[4788]: E1010 16:12:04.217918 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b\": container with ID starting with 081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b not found: ID does not exist" containerID="081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.217969 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b"} err="failed to get container status \"081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b\": rpc error: code = NotFound desc = could not find container \"081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b\": container with ID starting with 081ad69aefad338e6b2e6c52765b8067acc49b7ef5cba1aadd4127ff13732f5b not found: ID does not exist" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.217996 4788 scope.go:117] "RemoveContainer" containerID="cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa" Oct 10 16:12:04 crc kubenswrapper[4788]: E1010 16:12:04.218460 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa\": container with ID starting with cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa not found: ID does not exist" containerID="cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.218561 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa"} err="failed to get container status \"cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa\": rpc error: code = NotFound desc = could not find container \"cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa\": container with ID starting with cce907baddeb55fc1f56b53a4d0b8eb22c24464b10a6f454156af4bc42b82eaa not found: ID does not exist" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.315044 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.315204 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.315287 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5scg\" (UniqueName: \"kubernetes.io/projected/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-kube-api-access-k5scg\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.315308 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-config\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.315326 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-scripts\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.315679 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.316068 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-scripts\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.316738 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-config\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.320583 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.333599 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5scg\" (UniqueName: \"kubernetes.io/projected/511d54ed-e20b-4439-9a8f-ab1f738fb5c7-kube-api-access-k5scg\") pod \"ovn-northd-0\" (UID: \"511d54ed-e20b-4439-9a8f-ab1f738fb5c7\") " pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.386260 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.515589 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-92hfq"] Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.515708 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-92hfq"] Oct 10 16:12:04 crc kubenswrapper[4788]: I1010 16:12:04.834803 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 10 16:12:04 crc kubenswrapper[4788]: W1010 16:12:04.838299 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod511d54ed_e20b_4439_9a8f_ab1f738fb5c7.slice/crio-1c4cfd69e371c2b627d816c552add63f99b0e9beb3b396e4db67d7d12275f2b8 WatchSource:0}: Error finding container 1c4cfd69e371c2b627d816c552add63f99b0e9beb3b396e4db67d7d12275f2b8: Status 404 returned error can't find the container with id 1c4cfd69e371c2b627d816c552add63f99b0e9beb3b396e4db67d7d12275f2b8 Oct 10 16:12:05 crc kubenswrapper[4788]: I1010 16:12:05.166891 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"511d54ed-e20b-4439-9a8f-ab1f738fb5c7","Type":"ContainerStarted","Data":"9a013576f5389dc0888d11c308c39ea937b8baa0b2a244bac79e46f767d7b53a"} Oct 10 16:12:05 crc kubenswrapper[4788]: I1010 16:12:05.166952 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"511d54ed-e20b-4439-9a8f-ab1f738fb5c7","Type":"ContainerStarted","Data":"1c4cfd69e371c2b627d816c552add63f99b0e9beb3b396e4db67d7d12275f2b8"} Oct 10 16:12:06 crc kubenswrapper[4788]: I1010 16:12:06.176945 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"511d54ed-e20b-4439-9a8f-ab1f738fb5c7","Type":"ContainerStarted","Data":"3e00d4ec45747778de6f308c6d7f0011f038ad5766423063ecf5d92d93354b9b"} Oct 10 16:12:06 crc kubenswrapper[4788]: I1010 16:12:06.177489 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 10 16:12:06 crc kubenswrapper[4788]: I1010 16:12:06.208827 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.208797502 podStartE2EDuration="2.208797502s" podCreationTimestamp="2025-10-10 16:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:12:06.199704478 +0000 UTC m=+5228.649420046" watchObservedRunningTime="2025-10-10 16:12:06.208797502 +0000 UTC m=+5228.658513060" Oct 10 16:12:06 crc kubenswrapper[4788]: I1010 16:12:06.242535 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="436a3bba-9371-4bf0-938d-8655a4862e22" path="/var/lib/kubelet/pods/436a3bba-9371-4bf0-938d-8655a4862e22/volumes" Oct 10 16:12:09 crc kubenswrapper[4788]: I1010 16:12:09.836264 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-lx4nz"] Oct 10 16:12:09 crc kubenswrapper[4788]: I1010 16:12:09.838846 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lx4nz" Oct 10 16:12:09 crc kubenswrapper[4788]: I1010 16:12:09.847770 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lx4nz"] Oct 10 16:12:09 crc kubenswrapper[4788]: I1010 16:12:09.922186 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr7gg\" (UniqueName: \"kubernetes.io/projected/29e258a4-1339-4d09-9d34-bb75e08d761d-kube-api-access-hr7gg\") pod \"keystone-db-create-lx4nz\" (UID: \"29e258a4-1339-4d09-9d34-bb75e08d761d\") " pod="openstack/keystone-db-create-lx4nz" Oct 10 16:12:10 crc kubenswrapper[4788]: I1010 16:12:10.024100 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr7gg\" (UniqueName: \"kubernetes.io/projected/29e258a4-1339-4d09-9d34-bb75e08d761d-kube-api-access-hr7gg\") pod \"keystone-db-create-lx4nz\" (UID: \"29e258a4-1339-4d09-9d34-bb75e08d761d\") " pod="openstack/keystone-db-create-lx4nz" Oct 10 16:12:10 crc kubenswrapper[4788]: I1010 16:12:10.056993 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr7gg\" (UniqueName: \"kubernetes.io/projected/29e258a4-1339-4d09-9d34-bb75e08d761d-kube-api-access-hr7gg\") pod \"keystone-db-create-lx4nz\" (UID: \"29e258a4-1339-4d09-9d34-bb75e08d761d\") " pod="openstack/keystone-db-create-lx4nz" Oct 10 16:12:10 crc kubenswrapper[4788]: I1010 16:12:10.157651 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lx4nz" Oct 10 16:12:10 crc kubenswrapper[4788]: I1010 16:12:10.687617 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lx4nz"] Oct 10 16:12:11 crc kubenswrapper[4788]: I1010 16:12:11.250001 4788 generic.go:334] "Generic (PLEG): container finished" podID="29e258a4-1339-4d09-9d34-bb75e08d761d" containerID="52b5ef897cd24c50ed474f6ccd70bcc81d26072e3b46ec8e7117f788ace6386c" exitCode=0 Oct 10 16:12:11 crc kubenswrapper[4788]: I1010 16:12:11.250054 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lx4nz" event={"ID":"29e258a4-1339-4d09-9d34-bb75e08d761d","Type":"ContainerDied","Data":"52b5ef897cd24c50ed474f6ccd70bcc81d26072e3b46ec8e7117f788ace6386c"} Oct 10 16:12:11 crc kubenswrapper[4788]: I1010 16:12:11.250404 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lx4nz" event={"ID":"29e258a4-1339-4d09-9d34-bb75e08d761d","Type":"ContainerStarted","Data":"4a8fd747707fa34cd49162b2bef9200d3114b3fac62b3fefa0e2fd913be49b86"} Oct 10 16:12:12 crc kubenswrapper[4788]: I1010 16:12:12.202445 4788 scope.go:117] "RemoveContainer" containerID="a86a76c6c542a6e7ecc652116fd806df0529053d42ca71f427e59cb8c6fb743f" Oct 10 16:12:12 crc kubenswrapper[4788]: I1010 16:12:12.247977 4788 scope.go:117] "RemoveContainer" containerID="f4baf9d1df3290675f179a92f8a99c312666937fa9335670fa1f811d26854a81" Oct 10 16:12:12 crc kubenswrapper[4788]: I1010 16:12:12.274415 4788 scope.go:117] "RemoveContainer" containerID="8a05d46dcc24ec330f690eebc82e8cc99493efc3695dee962ea7324b8b8f55e3" Oct 10 16:12:12 crc kubenswrapper[4788]: I1010 16:12:12.300733 4788 scope.go:117] "RemoveContainer" containerID="03191736e94a84c761d81e80f55089678532cfe6a888f55d30a9b7081eca4003" Oct 10 16:12:12 crc kubenswrapper[4788]: I1010 16:12:12.564959 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lx4nz" Oct 10 16:12:12 crc kubenswrapper[4788]: I1010 16:12:12.681937 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr7gg\" (UniqueName: \"kubernetes.io/projected/29e258a4-1339-4d09-9d34-bb75e08d761d-kube-api-access-hr7gg\") pod \"29e258a4-1339-4d09-9d34-bb75e08d761d\" (UID: \"29e258a4-1339-4d09-9d34-bb75e08d761d\") " Oct 10 16:12:12 crc kubenswrapper[4788]: I1010 16:12:12.688961 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29e258a4-1339-4d09-9d34-bb75e08d761d-kube-api-access-hr7gg" (OuterVolumeSpecName: "kube-api-access-hr7gg") pod "29e258a4-1339-4d09-9d34-bb75e08d761d" (UID: "29e258a4-1339-4d09-9d34-bb75e08d761d"). InnerVolumeSpecName "kube-api-access-hr7gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:12 crc kubenswrapper[4788]: I1010 16:12:12.783921 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr7gg\" (UniqueName: \"kubernetes.io/projected/29e258a4-1339-4d09-9d34-bb75e08d761d-kube-api-access-hr7gg\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:13 crc kubenswrapper[4788]: I1010 16:12:13.274133 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lx4nz" event={"ID":"29e258a4-1339-4d09-9d34-bb75e08d761d","Type":"ContainerDied","Data":"4a8fd747707fa34cd49162b2bef9200d3114b3fac62b3fefa0e2fd913be49b86"} Oct 10 16:12:13 crc kubenswrapper[4788]: I1010 16:12:13.274209 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a8fd747707fa34cd49162b2bef9200d3114b3fac62b3fefa0e2fd913be49b86" Oct 10 16:12:13 crc kubenswrapper[4788]: I1010 16:12:13.274274 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lx4nz" Oct 10 16:12:17 crc kubenswrapper[4788]: I1010 16:12:17.753942 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kl957"] Oct 10 16:12:17 crc kubenswrapper[4788]: E1010 16:12:17.754964 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e258a4-1339-4d09-9d34-bb75e08d761d" containerName="mariadb-database-create" Oct 10 16:12:17 crc kubenswrapper[4788]: I1010 16:12:17.754984 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e258a4-1339-4d09-9d34-bb75e08d761d" containerName="mariadb-database-create" Oct 10 16:12:17 crc kubenswrapper[4788]: I1010 16:12:17.755278 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="29e258a4-1339-4d09-9d34-bb75e08d761d" containerName="mariadb-database-create" Oct 10 16:12:17 crc kubenswrapper[4788]: I1010 16:12:17.757039 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:17 crc kubenswrapper[4788]: I1010 16:12:17.761048 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kl957"] Oct 10 16:12:17 crc kubenswrapper[4788]: I1010 16:12:17.897917 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-catalog-content\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:17 crc kubenswrapper[4788]: I1010 16:12:17.897984 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-utilities\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:17 crc kubenswrapper[4788]: I1010 16:12:17.898249 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzr9n\" (UniqueName: \"kubernetes.io/projected/9d1145dc-82b9-4da5-8acf-50f6464e8347-kube-api-access-fzr9n\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:18 crc kubenswrapper[4788]: I1010 16:12:18.000180 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-catalog-content\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:18 crc kubenswrapper[4788]: I1010 16:12:18.000238 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-utilities\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:18 crc kubenswrapper[4788]: I1010 16:12:18.000338 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzr9n\" (UniqueName: \"kubernetes.io/projected/9d1145dc-82b9-4da5-8acf-50f6464e8347-kube-api-access-fzr9n\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:18 crc kubenswrapper[4788]: I1010 16:12:18.001070 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-catalog-content\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:18 crc kubenswrapper[4788]: I1010 16:12:18.001095 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-utilities\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:18 crc kubenswrapper[4788]: I1010 16:12:18.037460 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzr9n\" (UniqueName: \"kubernetes.io/projected/9d1145dc-82b9-4da5-8acf-50f6464e8347-kube-api-access-fzr9n\") pod \"certified-operators-kl957\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:18 crc kubenswrapper[4788]: I1010 16:12:18.085723 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:18 crc kubenswrapper[4788]: I1010 16:12:18.627085 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kl957"] Oct 10 16:12:18 crc kubenswrapper[4788]: W1010 16:12:18.635581 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d1145dc_82b9_4da5_8acf_50f6464e8347.slice/crio-c3e84c9b721a9b9d09dcf85c1b0da52e090c0d6668f9256a3ae07350bdb12bec WatchSource:0}: Error finding container c3e84c9b721a9b9d09dcf85c1b0da52e090c0d6668f9256a3ae07350bdb12bec: Status 404 returned error can't find the container with id c3e84c9b721a9b9d09dcf85c1b0da52e090c0d6668f9256a3ae07350bdb12bec Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.342948 4788 generic.go:334] "Generic (PLEG): container finished" podID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerID="8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3" exitCode=0 Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.343190 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kl957" event={"ID":"9d1145dc-82b9-4da5-8acf-50f6464e8347","Type":"ContainerDied","Data":"8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3"} Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.343838 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kl957" event={"ID":"9d1145dc-82b9-4da5-8acf-50f6464e8347","Type":"ContainerStarted","Data":"c3e84c9b721a9b9d09dcf85c1b0da52e090c0d6668f9256a3ae07350bdb12bec"} Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.346800 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.477102 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.806419 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ec63-account-create-9gfbp"] Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.807675 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ec63-account-create-9gfbp" Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.809304 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.818648 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ec63-account-create-9gfbp"] Oct 10 16:12:19 crc kubenswrapper[4788]: I1010 16:12:19.936781 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nms2h\" (UniqueName: \"kubernetes.io/projected/81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23-kube-api-access-nms2h\") pod \"keystone-ec63-account-create-9gfbp\" (UID: \"81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23\") " pod="openstack/keystone-ec63-account-create-9gfbp" Oct 10 16:12:20 crc kubenswrapper[4788]: I1010 16:12:20.037941 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nms2h\" (UniqueName: \"kubernetes.io/projected/81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23-kube-api-access-nms2h\") pod \"keystone-ec63-account-create-9gfbp\" (UID: \"81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23\") " pod="openstack/keystone-ec63-account-create-9gfbp" Oct 10 16:12:20 crc kubenswrapper[4788]: I1010 16:12:20.059787 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nms2h\" (UniqueName: \"kubernetes.io/projected/81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23-kube-api-access-nms2h\") pod \"keystone-ec63-account-create-9gfbp\" (UID: \"81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23\") " pod="openstack/keystone-ec63-account-create-9gfbp" Oct 10 16:12:20 crc kubenswrapper[4788]: I1010 16:12:20.124358 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ec63-account-create-9gfbp" Oct 10 16:12:20 crc kubenswrapper[4788]: I1010 16:12:20.358730 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kl957" event={"ID":"9d1145dc-82b9-4da5-8acf-50f6464e8347","Type":"ContainerStarted","Data":"7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030"} Oct 10 16:12:20 crc kubenswrapper[4788]: I1010 16:12:20.569499 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ec63-account-create-9gfbp"] Oct 10 16:12:20 crc kubenswrapper[4788]: W1010 16:12:20.635517 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81e2b2a4_1ee3_4164_8f80_b1bf5ebdca23.slice/crio-0231bcbac81f9ece6cce1da09a4c1b069cfa38ba7d50735e03ce339ddd047804 WatchSource:0}: Error finding container 0231bcbac81f9ece6cce1da09a4c1b069cfa38ba7d50735e03ce339ddd047804: Status 404 returned error can't find the container with id 0231bcbac81f9ece6cce1da09a4c1b069cfa38ba7d50735e03ce339ddd047804 Oct 10 16:12:21 crc kubenswrapper[4788]: I1010 16:12:21.373850 4788 generic.go:334] "Generic (PLEG): container finished" podID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerID="7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030" exitCode=0 Oct 10 16:12:21 crc kubenswrapper[4788]: I1010 16:12:21.374003 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kl957" event={"ID":"9d1145dc-82b9-4da5-8acf-50f6464e8347","Type":"ContainerDied","Data":"7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030"} Oct 10 16:12:21 crc kubenswrapper[4788]: I1010 16:12:21.376723 4788 generic.go:334] "Generic (PLEG): container finished" podID="81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23" containerID="7564209e7512c76e32610dafec7a18b7d2b77184904e34abbd361ab7cbab0884" exitCode=0 Oct 10 16:12:21 crc kubenswrapper[4788]: I1010 16:12:21.376772 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ec63-account-create-9gfbp" event={"ID":"81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23","Type":"ContainerDied","Data":"7564209e7512c76e32610dafec7a18b7d2b77184904e34abbd361ab7cbab0884"} Oct 10 16:12:21 crc kubenswrapper[4788]: I1010 16:12:21.376831 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ec63-account-create-9gfbp" event={"ID":"81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23","Type":"ContainerStarted","Data":"0231bcbac81f9ece6cce1da09a4c1b069cfa38ba7d50735e03ce339ddd047804"} Oct 10 16:12:22 crc kubenswrapper[4788]: I1010 16:12:22.389931 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kl957" event={"ID":"9d1145dc-82b9-4da5-8acf-50f6464e8347","Type":"ContainerStarted","Data":"44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2"} Oct 10 16:12:22 crc kubenswrapper[4788]: I1010 16:12:22.413106 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kl957" podStartSLOduration=2.58291238 podStartE2EDuration="5.413090836s" podCreationTimestamp="2025-10-10 16:12:17 +0000 UTC" firstStartedPulling="2025-10-10 16:12:19.346499597 +0000 UTC m=+5241.796215145" lastFinishedPulling="2025-10-10 16:12:22.176678043 +0000 UTC m=+5244.626393601" observedRunningTime="2025-10-10 16:12:22.41013098 +0000 UTC m=+5244.859846528" watchObservedRunningTime="2025-10-10 16:12:22.413090836 +0000 UTC m=+5244.862806384" Oct 10 16:12:22 crc kubenswrapper[4788]: I1010 16:12:22.757428 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ec63-account-create-9gfbp" Oct 10 16:12:22 crc kubenswrapper[4788]: I1010 16:12:22.888972 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nms2h\" (UniqueName: \"kubernetes.io/projected/81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23-kube-api-access-nms2h\") pod \"81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23\" (UID: \"81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23\") " Oct 10 16:12:22 crc kubenswrapper[4788]: I1010 16:12:22.898415 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23-kube-api-access-nms2h" (OuterVolumeSpecName: "kube-api-access-nms2h") pod "81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23" (UID: "81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23"). InnerVolumeSpecName "kube-api-access-nms2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:22 crc kubenswrapper[4788]: I1010 16:12:22.991509 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nms2h\" (UniqueName: \"kubernetes.io/projected/81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23-kube-api-access-nms2h\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:23 crc kubenswrapper[4788]: I1010 16:12:23.403518 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ec63-account-create-9gfbp" event={"ID":"81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23","Type":"ContainerDied","Data":"0231bcbac81f9ece6cce1da09a4c1b069cfa38ba7d50735e03ce339ddd047804"} Oct 10 16:12:23 crc kubenswrapper[4788]: I1010 16:12:23.404452 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0231bcbac81f9ece6cce1da09a4c1b069cfa38ba7d50735e03ce339ddd047804" Oct 10 16:12:23 crc kubenswrapper[4788]: I1010 16:12:23.403586 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ec63-account-create-9gfbp" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.295502 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xbzjx"] Oct 10 16:12:25 crc kubenswrapper[4788]: E1010 16:12:25.296220 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23" containerName="mariadb-account-create" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.296236 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23" containerName="mariadb-account-create" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.296427 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23" containerName="mariadb-account-create" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.297047 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.300602 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rvj9h" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.300790 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.301711 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.305768 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.316655 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xbzjx"] Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.433790 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-combined-ca-bundle\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.433893 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-config-data\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.433921 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms8dj\" (UniqueName: \"kubernetes.io/projected/459619f5-5e06-467c-b778-3b063e727da4-kube-api-access-ms8dj\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.536302 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-combined-ca-bundle\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.536412 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-config-data\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.536444 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms8dj\" (UniqueName: \"kubernetes.io/projected/459619f5-5e06-467c-b778-3b063e727da4-kube-api-access-ms8dj\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.542835 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-combined-ca-bundle\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.543829 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-config-data\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.562816 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms8dj\" (UniqueName: \"kubernetes.io/projected/459619f5-5e06-467c-b778-3b063e727da4-kube-api-access-ms8dj\") pod \"keystone-db-sync-xbzjx\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:25 crc kubenswrapper[4788]: I1010 16:12:25.618365 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:26 crc kubenswrapper[4788]: I1010 16:12:26.053441 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xbzjx"] Oct 10 16:12:26 crc kubenswrapper[4788]: I1010 16:12:26.428590 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xbzjx" event={"ID":"459619f5-5e06-467c-b778-3b063e727da4","Type":"ContainerStarted","Data":"0e3848714678ca88f78feeaef5b60ab7e762fab156a35b1fdff6b8624884102c"} Oct 10 16:12:26 crc kubenswrapper[4788]: I1010 16:12:26.428915 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xbzjx" event={"ID":"459619f5-5e06-467c-b778-3b063e727da4","Type":"ContainerStarted","Data":"b45a4ef090a387c8ce4bcb793fde60270fd3ec9b33cfc86123d4f25e0098762d"} Oct 10 16:12:26 crc kubenswrapper[4788]: I1010 16:12:26.466285 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xbzjx" podStartSLOduration=1.466261801 podStartE2EDuration="1.466261801s" podCreationTimestamp="2025-10-10 16:12:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:12:26.452401975 +0000 UTC m=+5248.902117543" watchObservedRunningTime="2025-10-10 16:12:26.466261801 +0000 UTC m=+5248.915977349" Oct 10 16:12:28 crc kubenswrapper[4788]: I1010 16:12:28.086055 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:28 crc kubenswrapper[4788]: I1010 16:12:28.086449 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:28 crc kubenswrapper[4788]: I1010 16:12:28.136794 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:28 crc kubenswrapper[4788]: I1010 16:12:28.444880 4788 generic.go:334] "Generic (PLEG): container finished" podID="459619f5-5e06-467c-b778-3b063e727da4" containerID="0e3848714678ca88f78feeaef5b60ab7e762fab156a35b1fdff6b8624884102c" exitCode=0 Oct 10 16:12:28 crc kubenswrapper[4788]: I1010 16:12:28.444968 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xbzjx" event={"ID":"459619f5-5e06-467c-b778-3b063e727da4","Type":"ContainerDied","Data":"0e3848714678ca88f78feeaef5b60ab7e762fab156a35b1fdff6b8624884102c"} Oct 10 16:12:28 crc kubenswrapper[4788]: I1010 16:12:28.495781 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:29 crc kubenswrapper[4788]: I1010 16:12:29.782294 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:29 crc kubenswrapper[4788]: I1010 16:12:29.927473 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms8dj\" (UniqueName: \"kubernetes.io/projected/459619f5-5e06-467c-b778-3b063e727da4-kube-api-access-ms8dj\") pod \"459619f5-5e06-467c-b778-3b063e727da4\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " Oct 10 16:12:29 crc kubenswrapper[4788]: I1010 16:12:29.927599 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-config-data\") pod \"459619f5-5e06-467c-b778-3b063e727da4\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " Oct 10 16:12:29 crc kubenswrapper[4788]: I1010 16:12:29.927646 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-combined-ca-bundle\") pod \"459619f5-5e06-467c-b778-3b063e727da4\" (UID: \"459619f5-5e06-467c-b778-3b063e727da4\") " Oct 10 16:12:29 crc kubenswrapper[4788]: I1010 16:12:29.936622 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/459619f5-5e06-467c-b778-3b063e727da4-kube-api-access-ms8dj" (OuterVolumeSpecName: "kube-api-access-ms8dj") pod "459619f5-5e06-467c-b778-3b063e727da4" (UID: "459619f5-5e06-467c-b778-3b063e727da4"). InnerVolumeSpecName "kube-api-access-ms8dj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:29 crc kubenswrapper[4788]: I1010 16:12:29.954466 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "459619f5-5e06-467c-b778-3b063e727da4" (UID: "459619f5-5e06-467c-b778-3b063e727da4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:29 crc kubenswrapper[4788]: I1010 16:12:29.983269 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-config-data" (OuterVolumeSpecName: "config-data") pod "459619f5-5e06-467c-b778-3b063e727da4" (UID: "459619f5-5e06-467c-b778-3b063e727da4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:30 crc kubenswrapper[4788]: I1010 16:12:30.037110 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms8dj\" (UniqueName: \"kubernetes.io/projected/459619f5-5e06-467c-b778-3b063e727da4-kube-api-access-ms8dj\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:30 crc kubenswrapper[4788]: I1010 16:12:30.037444 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:30 crc kubenswrapper[4788]: I1010 16:12:30.037454 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/459619f5-5e06-467c-b778-3b063e727da4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:30 crc kubenswrapper[4788]: I1010 16:12:30.469032 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xbzjx" event={"ID":"459619f5-5e06-467c-b778-3b063e727da4","Type":"ContainerDied","Data":"b45a4ef090a387c8ce4bcb793fde60270fd3ec9b33cfc86123d4f25e0098762d"} Oct 10 16:12:30 crc kubenswrapper[4788]: I1010 16:12:30.469080 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b45a4ef090a387c8ce4bcb793fde60270fd3ec9b33cfc86123d4f25e0098762d" Oct 10 16:12:30 crc kubenswrapper[4788]: I1010 16:12:30.469170 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xbzjx" Oct 10 16:12:30 crc kubenswrapper[4788]: I1010 16:12:30.533614 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kl957"] Oct 10 16:12:30 crc kubenswrapper[4788]: I1010 16:12:30.534054 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kl957" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerName="registry-server" containerID="cri-o://44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2" gracePeriod=2 Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.002860 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.089056 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-99458d6fc-ngrpb"] Oct 10 16:12:31 crc kubenswrapper[4788]: E1010 16:12:31.089538 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerName="extract-content" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.089557 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerName="extract-content" Oct 10 16:12:31 crc kubenswrapper[4788]: E1010 16:12:31.089645 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerName="extract-utilities" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.089659 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerName="extract-utilities" Oct 10 16:12:31 crc kubenswrapper[4788]: E1010 16:12:31.089676 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="459619f5-5e06-467c-b778-3b063e727da4" containerName="keystone-db-sync" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.089683 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="459619f5-5e06-467c-b778-3b063e727da4" containerName="keystone-db-sync" Oct 10 16:12:31 crc kubenswrapper[4788]: E1010 16:12:31.089714 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerName="registry-server" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.089721 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerName="registry-server" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.089878 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="459619f5-5e06-467c-b778-3b063e727da4" containerName="keystone-db-sync" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.089900 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerName="registry-server" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.091183 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.097743 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-99458d6fc-ngrpb"] Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.106637 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-pm6fd"] Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.107787 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.110730 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.112202 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.112819 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rvj9h" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.117901 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.120613 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pm6fd"] Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.170526 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-utilities\") pod \"9d1145dc-82b9-4da5-8acf-50f6464e8347\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.170699 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-catalog-content\") pod \"9d1145dc-82b9-4da5-8acf-50f6464e8347\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.170720 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzr9n\" (UniqueName: \"kubernetes.io/projected/9d1145dc-82b9-4da5-8acf-50f6464e8347-kube-api-access-fzr9n\") pod \"9d1145dc-82b9-4da5-8acf-50f6464e8347\" (UID: \"9d1145dc-82b9-4da5-8acf-50f6464e8347\") " Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.170941 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-sb\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.171015 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-config\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.171057 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-dns-svc\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.171081 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-nb\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.171097 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj78w\" (UniqueName: \"kubernetes.io/projected/ef832418-6747-45e3-8a6f-4f07eb2b5107-kube-api-access-zj78w\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.171556 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-utilities" (OuterVolumeSpecName: "utilities") pod "9d1145dc-82b9-4da5-8acf-50f6464e8347" (UID: "9d1145dc-82b9-4da5-8acf-50f6464e8347"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.179316 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d1145dc-82b9-4da5-8acf-50f6464e8347-kube-api-access-fzr9n" (OuterVolumeSpecName: "kube-api-access-fzr9n") pod "9d1145dc-82b9-4da5-8acf-50f6464e8347" (UID: "9d1145dc-82b9-4da5-8acf-50f6464e8347"). InnerVolumeSpecName "kube-api-access-fzr9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.216889 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d1145dc-82b9-4da5-8acf-50f6464e8347" (UID: "9d1145dc-82b9-4da5-8acf-50f6464e8347"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.272643 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-config\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.272711 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-scripts\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.272733 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-config-data\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.272753 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-fernet-keys\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.272768 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-combined-ca-bundle\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.272793 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-dns-svc\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.272907 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-nb\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.272945 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj78w\" (UniqueName: \"kubernetes.io/projected/ef832418-6747-45e3-8a6f-4f07eb2b5107-kube-api-access-zj78w\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.273069 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-credential-keys\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.273111 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-sb\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.273170 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbqc9\" (UniqueName: \"kubernetes.io/projected/fb689023-f357-447e-befd-da1a8eaa0287-kube-api-access-qbqc9\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.273408 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.273426 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d1145dc-82b9-4da5-8acf-50f6464e8347-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.273439 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzr9n\" (UniqueName: \"kubernetes.io/projected/9d1145dc-82b9-4da5-8acf-50f6464e8347-kube-api-access-fzr9n\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.273688 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-dns-svc\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.273773 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-config\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.274345 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-nb\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.274365 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-sb\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.294660 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj78w\" (UniqueName: \"kubernetes.io/projected/ef832418-6747-45e3-8a6f-4f07eb2b5107-kube-api-access-zj78w\") pod \"dnsmasq-dns-99458d6fc-ngrpb\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.375307 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-scripts\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.375351 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-config-data\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.375369 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-fernet-keys\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.375386 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-combined-ca-bundle\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.375442 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-credential-keys\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.375489 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbqc9\" (UniqueName: \"kubernetes.io/projected/fb689023-f357-447e-befd-da1a8eaa0287-kube-api-access-qbqc9\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.379165 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-combined-ca-bundle\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.379742 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-config-data\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.380018 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-scripts\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.380456 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-credential-keys\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.380850 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-fernet-keys\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.400907 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbqc9\" (UniqueName: \"kubernetes.io/projected/fb689023-f357-447e-befd-da1a8eaa0287-kube-api-access-qbqc9\") pod \"keystone-bootstrap-pm6fd\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.419220 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.436658 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.480871 4788 generic.go:334] "Generic (PLEG): container finished" podID="9d1145dc-82b9-4da5-8acf-50f6464e8347" containerID="44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2" exitCode=0 Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.480909 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kl957" event={"ID":"9d1145dc-82b9-4da5-8acf-50f6464e8347","Type":"ContainerDied","Data":"44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2"} Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.480938 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kl957" event={"ID":"9d1145dc-82b9-4da5-8acf-50f6464e8347","Type":"ContainerDied","Data":"c3e84c9b721a9b9d09dcf85c1b0da52e090c0d6668f9256a3ae07350bdb12bec"} Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.480955 4788 scope.go:117] "RemoveContainer" containerID="44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.480988 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kl957" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.520011 4788 scope.go:117] "RemoveContainer" containerID="7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.524044 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kl957"] Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.532251 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kl957"] Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.592789 4788 scope.go:117] "RemoveContainer" containerID="8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.674208 4788 scope.go:117] "RemoveContainer" containerID="44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2" Oct 10 16:12:31 crc kubenswrapper[4788]: E1010 16:12:31.674935 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2\": container with ID starting with 44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2 not found: ID does not exist" containerID="44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.674977 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2"} err="failed to get container status \"44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2\": rpc error: code = NotFound desc = could not find container \"44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2\": container with ID starting with 44d2d31391fc47ace7c8b9096a8a2da14288a885e92be175f4d39a48ba030af2 not found: ID does not exist" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.675006 4788 scope.go:117] "RemoveContainer" containerID="7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030" Oct 10 16:12:31 crc kubenswrapper[4788]: E1010 16:12:31.675287 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030\": container with ID starting with 7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030 not found: ID does not exist" containerID="7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.675305 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030"} err="failed to get container status \"7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030\": rpc error: code = NotFound desc = could not find container \"7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030\": container with ID starting with 7fb3c773ff823bf035cee2fc1b241b285b4f84d8124784b1dafd977207dc2030 not found: ID does not exist" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.675318 4788 scope.go:117] "RemoveContainer" containerID="8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3" Oct 10 16:12:31 crc kubenswrapper[4788]: E1010 16:12:31.676088 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3\": container with ID starting with 8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3 not found: ID does not exist" containerID="8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.676128 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3"} err="failed to get container status \"8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3\": rpc error: code = NotFound desc = could not find container \"8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3\": container with ID starting with 8d0f61f0d8600077e2ee46723ed19e8e344fffbde5bfce8cbc75aa87378108c3 not found: ID does not exist" Oct 10 16:12:31 crc kubenswrapper[4788]: I1010 16:12:31.931274 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-99458d6fc-ngrpb"] Oct 10 16:12:32 crc kubenswrapper[4788]: I1010 16:12:32.031850 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pm6fd"] Oct 10 16:12:32 crc kubenswrapper[4788]: W1010 16:12:32.043248 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb689023_f357_447e_befd_da1a8eaa0287.slice/crio-273f47166466a814e4b36e626977ba4887afd91ffa25ad4e071769603a7ae856 WatchSource:0}: Error finding container 273f47166466a814e4b36e626977ba4887afd91ffa25ad4e071769603a7ae856: Status 404 returned error can't find the container with id 273f47166466a814e4b36e626977ba4887afd91ffa25ad4e071769603a7ae856 Oct 10 16:12:32 crc kubenswrapper[4788]: I1010 16:12:32.246940 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d1145dc-82b9-4da5-8acf-50f6464e8347" path="/var/lib/kubelet/pods/9d1145dc-82b9-4da5-8acf-50f6464e8347/volumes" Oct 10 16:12:32 crc kubenswrapper[4788]: I1010 16:12:32.490626 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pm6fd" event={"ID":"fb689023-f357-447e-befd-da1a8eaa0287","Type":"ContainerStarted","Data":"1c6717185a7b9e9ecdd28f9360f13a808cd80aa46ad52be555766c9af8df9ac1"} Oct 10 16:12:32 crc kubenswrapper[4788]: I1010 16:12:32.490670 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pm6fd" event={"ID":"fb689023-f357-447e-befd-da1a8eaa0287","Type":"ContainerStarted","Data":"273f47166466a814e4b36e626977ba4887afd91ffa25ad4e071769603a7ae856"} Oct 10 16:12:32 crc kubenswrapper[4788]: I1010 16:12:32.492717 4788 generic.go:334] "Generic (PLEG): container finished" podID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerID="322a9cdb7fd931e27e56a653ac4d0d77311acc7367d9433d880454a178ec8c01" exitCode=0 Oct 10 16:12:32 crc kubenswrapper[4788]: I1010 16:12:32.492752 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" event={"ID":"ef832418-6747-45e3-8a6f-4f07eb2b5107","Type":"ContainerDied","Data":"322a9cdb7fd931e27e56a653ac4d0d77311acc7367d9433d880454a178ec8c01"} Oct 10 16:12:32 crc kubenswrapper[4788]: I1010 16:12:32.492768 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" event={"ID":"ef832418-6747-45e3-8a6f-4f07eb2b5107","Type":"ContainerStarted","Data":"f6f2290a2170eea648e9a3b32365e920a616996efc4d87ce503b02cd8ee51566"} Oct 10 16:12:32 crc kubenswrapper[4788]: I1010 16:12:32.513557 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-pm6fd" podStartSLOduration=1.513526224 podStartE2EDuration="1.513526224s" podCreationTimestamp="2025-10-10 16:12:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:12:32.510966038 +0000 UTC m=+5254.960681586" watchObservedRunningTime="2025-10-10 16:12:32.513526224 +0000 UTC m=+5254.963241812" Oct 10 16:12:33 crc kubenswrapper[4788]: I1010 16:12:33.505578 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" event={"ID":"ef832418-6747-45e3-8a6f-4f07eb2b5107","Type":"ContainerStarted","Data":"91022de13ed6443c1cd593c5d52d8ac0f587968ce49cc5d9078bb3eaeb741f40"} Oct 10 16:12:33 crc kubenswrapper[4788]: I1010 16:12:33.548289 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" podStartSLOduration=2.548269206 podStartE2EDuration="2.548269206s" podCreationTimestamp="2025-10-10 16:12:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:12:33.537676245 +0000 UTC m=+5255.987391813" watchObservedRunningTime="2025-10-10 16:12:33.548269206 +0000 UTC m=+5255.997984754" Oct 10 16:12:34 crc kubenswrapper[4788]: I1010 16:12:34.516113 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.335020 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h69gb"] Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.337709 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.357839 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h69gb"] Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.491395 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rtms\" (UniqueName: \"kubernetes.io/projected/a27976f0-d788-4750-be52-148b419d7e9c-kube-api-access-9rtms\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.491533 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-utilities\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.491604 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-catalog-content\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.525702 4788 generic.go:334] "Generic (PLEG): container finished" podID="fb689023-f357-447e-befd-da1a8eaa0287" containerID="1c6717185a7b9e9ecdd28f9360f13a808cd80aa46ad52be555766c9af8df9ac1" exitCode=0 Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.525798 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pm6fd" event={"ID":"fb689023-f357-447e-befd-da1a8eaa0287","Type":"ContainerDied","Data":"1c6717185a7b9e9ecdd28f9360f13a808cd80aa46ad52be555766c9af8df9ac1"} Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.592892 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-utilities\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.593009 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-catalog-content\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.593046 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rtms\" (UniqueName: \"kubernetes.io/projected/a27976f0-d788-4750-be52-148b419d7e9c-kube-api-access-9rtms\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.593460 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-utilities\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.593566 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-catalog-content\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.614924 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rtms\" (UniqueName: \"kubernetes.io/projected/a27976f0-d788-4750-be52-148b419d7e9c-kube-api-access-9rtms\") pod \"redhat-operators-h69gb\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:35 crc kubenswrapper[4788]: I1010 16:12:35.661246 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:36 crc kubenswrapper[4788]: I1010 16:12:36.125868 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h69gb"] Oct 10 16:12:36 crc kubenswrapper[4788]: I1010 16:12:36.533353 4788 generic.go:334] "Generic (PLEG): container finished" podID="a27976f0-d788-4750-be52-148b419d7e9c" containerID="1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9" exitCode=0 Oct 10 16:12:36 crc kubenswrapper[4788]: I1010 16:12:36.533441 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69gb" event={"ID":"a27976f0-d788-4750-be52-148b419d7e9c","Type":"ContainerDied","Data":"1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9"} Oct 10 16:12:36 crc kubenswrapper[4788]: I1010 16:12:36.533679 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69gb" event={"ID":"a27976f0-d788-4750-be52-148b419d7e9c","Type":"ContainerStarted","Data":"aeb1d9ace0d8fb92e3c4edf459bea53ef7637d977bb5492c63d8facc24e41d8d"} Oct 10 16:12:36 crc kubenswrapper[4788]: I1010 16:12:36.837965 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.017098 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-fernet-keys\") pod \"fb689023-f357-447e-befd-da1a8eaa0287\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.017185 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbqc9\" (UniqueName: \"kubernetes.io/projected/fb689023-f357-447e-befd-da1a8eaa0287-kube-api-access-qbqc9\") pod \"fb689023-f357-447e-befd-da1a8eaa0287\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.017339 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-combined-ca-bundle\") pod \"fb689023-f357-447e-befd-da1a8eaa0287\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.017411 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-scripts\") pod \"fb689023-f357-447e-befd-da1a8eaa0287\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.017458 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-config-data\") pod \"fb689023-f357-447e-befd-da1a8eaa0287\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.017499 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-credential-keys\") pod \"fb689023-f357-447e-befd-da1a8eaa0287\" (UID: \"fb689023-f357-447e-befd-da1a8eaa0287\") " Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.023015 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fb689023-f357-447e-befd-da1a8eaa0287" (UID: "fb689023-f357-447e-befd-da1a8eaa0287"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.023067 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fb689023-f357-447e-befd-da1a8eaa0287" (UID: "fb689023-f357-447e-befd-da1a8eaa0287"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.024308 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-scripts" (OuterVolumeSpecName: "scripts") pod "fb689023-f357-447e-befd-da1a8eaa0287" (UID: "fb689023-f357-447e-befd-da1a8eaa0287"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.024386 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb689023-f357-447e-befd-da1a8eaa0287-kube-api-access-qbqc9" (OuterVolumeSpecName: "kube-api-access-qbqc9") pod "fb689023-f357-447e-befd-da1a8eaa0287" (UID: "fb689023-f357-447e-befd-da1a8eaa0287"). InnerVolumeSpecName "kube-api-access-qbqc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.041405 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb689023-f357-447e-befd-da1a8eaa0287" (UID: "fb689023-f357-447e-befd-da1a8eaa0287"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.042351 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-config-data" (OuterVolumeSpecName: "config-data") pod "fb689023-f357-447e-befd-da1a8eaa0287" (UID: "fb689023-f357-447e-befd-da1a8eaa0287"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.119710 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.119765 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.119824 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.119835 4788 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.119845 4788 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fb689023-f357-447e-befd-da1a8eaa0287-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.119854 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbqc9\" (UniqueName: \"kubernetes.io/projected/fb689023-f357-447e-befd-da1a8eaa0287-kube-api-access-qbqc9\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.552801 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pm6fd" event={"ID":"fb689023-f357-447e-befd-da1a8eaa0287","Type":"ContainerDied","Data":"273f47166466a814e4b36e626977ba4887afd91ffa25ad4e071769603a7ae856"} Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.552837 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pm6fd" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.552846 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="273f47166466a814e4b36e626977ba4887afd91ffa25ad4e071769603a7ae856" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.555952 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69gb" event={"ID":"a27976f0-d788-4750-be52-148b419d7e9c","Type":"ContainerStarted","Data":"a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb"} Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.633275 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-pm6fd"] Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.641745 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-pm6fd"] Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.719951 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sb82b"] Oct 10 16:12:37 crc kubenswrapper[4788]: E1010 16:12:37.720355 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb689023-f357-447e-befd-da1a8eaa0287" containerName="keystone-bootstrap" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.720372 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb689023-f357-447e-befd-da1a8eaa0287" containerName="keystone-bootstrap" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.720533 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb689023-f357-447e-befd-da1a8eaa0287" containerName="keystone-bootstrap" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.721123 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.722844 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.722961 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rvj9h" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.729585 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.729883 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.743449 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sb82b"] Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.832916 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-scripts\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.833000 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-credential-keys\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.833048 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-fernet-keys\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.833193 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-combined-ca-bundle\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.833470 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bxzq\" (UniqueName: \"kubernetes.io/projected/8bf92e46-caea-4151-852c-7e29c99bb230-kube-api-access-7bxzq\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.833867 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-config-data\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.935603 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-config-data\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.935668 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-scripts\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.935702 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-credential-keys\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.935735 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-fernet-keys\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.935821 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-combined-ca-bundle\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.935880 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bxzq\" (UniqueName: \"kubernetes.io/projected/8bf92e46-caea-4151-852c-7e29c99bb230-kube-api-access-7bxzq\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.939559 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-scripts\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.940001 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-fernet-keys\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.940221 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-config-data\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.940396 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-credential-keys\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.940750 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-combined-ca-bundle\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:37 crc kubenswrapper[4788]: I1010 16:12:37.955087 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bxzq\" (UniqueName: \"kubernetes.io/projected/8bf92e46-caea-4151-852c-7e29c99bb230-kube-api-access-7bxzq\") pod \"keystone-bootstrap-sb82b\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:38 crc kubenswrapper[4788]: I1010 16:12:38.086908 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:38 crc kubenswrapper[4788]: I1010 16:12:38.242742 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb689023-f357-447e-befd-da1a8eaa0287" path="/var/lib/kubelet/pods/fb689023-f357-447e-befd-da1a8eaa0287/volumes" Oct 10 16:12:38 crc kubenswrapper[4788]: I1010 16:12:38.352853 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sb82b"] Oct 10 16:12:38 crc kubenswrapper[4788]: W1010 16:12:38.353439 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bf92e46_caea_4151_852c_7e29c99bb230.slice/crio-e142f8746bf7406b8d897f01f9eb02e4415feb99e4f25e68330ce82109592454 WatchSource:0}: Error finding container e142f8746bf7406b8d897f01f9eb02e4415feb99e4f25e68330ce82109592454: Status 404 returned error can't find the container with id e142f8746bf7406b8d897f01f9eb02e4415feb99e4f25e68330ce82109592454 Oct 10 16:12:38 crc kubenswrapper[4788]: I1010 16:12:38.565267 4788 generic.go:334] "Generic (PLEG): container finished" podID="a27976f0-d788-4750-be52-148b419d7e9c" containerID="a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb" exitCode=0 Oct 10 16:12:38 crc kubenswrapper[4788]: I1010 16:12:38.565437 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69gb" event={"ID":"a27976f0-d788-4750-be52-148b419d7e9c","Type":"ContainerDied","Data":"a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb"} Oct 10 16:12:38 crc kubenswrapper[4788]: I1010 16:12:38.567808 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sb82b" event={"ID":"8bf92e46-caea-4151-852c-7e29c99bb230","Type":"ContainerStarted","Data":"2e8184017a6d3b1f51d23bc5e9b714a36c72a832e02816022fed273a2eabb69d"} Oct 10 16:12:38 crc kubenswrapper[4788]: I1010 16:12:38.567872 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sb82b" event={"ID":"8bf92e46-caea-4151-852c-7e29c99bb230","Type":"ContainerStarted","Data":"e142f8746bf7406b8d897f01f9eb02e4415feb99e4f25e68330ce82109592454"} Oct 10 16:12:38 crc kubenswrapper[4788]: I1010 16:12:38.607559 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sb82b" podStartSLOduration=1.607539828 podStartE2EDuration="1.607539828s" podCreationTimestamp="2025-10-10 16:12:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:12:38.597701855 +0000 UTC m=+5261.047417403" watchObservedRunningTime="2025-10-10 16:12:38.607539828 +0000 UTC m=+5261.057255386" Oct 10 16:12:39 crc kubenswrapper[4788]: I1010 16:12:39.578075 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69gb" event={"ID":"a27976f0-d788-4750-be52-148b419d7e9c","Type":"ContainerStarted","Data":"40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f"} Oct 10 16:12:39 crc kubenswrapper[4788]: I1010 16:12:39.603088 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h69gb" podStartSLOduration=1.858293129 podStartE2EDuration="4.603064493s" podCreationTimestamp="2025-10-10 16:12:35 +0000 UTC" firstStartedPulling="2025-10-10 16:12:36.535270701 +0000 UTC m=+5258.984986249" lastFinishedPulling="2025-10-10 16:12:39.280042035 +0000 UTC m=+5261.729757613" observedRunningTime="2025-10-10 16:12:39.598028794 +0000 UTC m=+5262.047744362" watchObservedRunningTime="2025-10-10 16:12:39.603064493 +0000 UTC m=+5262.052780061" Oct 10 16:12:41 crc kubenswrapper[4788]: I1010 16:12:41.421778 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:12:41 crc kubenswrapper[4788]: I1010 16:12:41.471956 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-ccf97d949-s6wcz"] Oct 10 16:12:41 crc kubenswrapper[4788]: I1010 16:12:41.472197 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" podUID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" containerName="dnsmasq-dns" containerID="cri-o://227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0" gracePeriod=10 Oct 10 16:12:41 crc kubenswrapper[4788]: I1010 16:12:41.597007 4788 generic.go:334] "Generic (PLEG): container finished" podID="8bf92e46-caea-4151-852c-7e29c99bb230" containerID="2e8184017a6d3b1f51d23bc5e9b714a36c72a832e02816022fed273a2eabb69d" exitCode=0 Oct 10 16:12:41 crc kubenswrapper[4788]: I1010 16:12:41.597100 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sb82b" event={"ID":"8bf92e46-caea-4151-852c-7e29c99bb230","Type":"ContainerDied","Data":"2e8184017a6d3b1f51d23bc5e9b714a36c72a832e02816022fed273a2eabb69d"} Oct 10 16:12:41 crc kubenswrapper[4788]: I1010 16:12:41.980854 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.110553 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-config\") pod \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.111345 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-dns-svc\") pod \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.111446 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-sb\") pod \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.111473 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b7p5\" (UniqueName: \"kubernetes.io/projected/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-kube-api-access-8b7p5\") pod \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.111498 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-nb\") pod \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\" (UID: \"ae76828a-d10f-49d2-94ae-39c6cd67ea0c\") " Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.116127 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-kube-api-access-8b7p5" (OuterVolumeSpecName: "kube-api-access-8b7p5") pod "ae76828a-d10f-49d2-94ae-39c6cd67ea0c" (UID: "ae76828a-d10f-49d2-94ae-39c6cd67ea0c"). InnerVolumeSpecName "kube-api-access-8b7p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.150342 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae76828a-d10f-49d2-94ae-39c6cd67ea0c" (UID: "ae76828a-d10f-49d2-94ae-39c6cd67ea0c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.153311 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ae76828a-d10f-49d2-94ae-39c6cd67ea0c" (UID: "ae76828a-d10f-49d2-94ae-39c6cd67ea0c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.153812 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-config" (OuterVolumeSpecName: "config") pod "ae76828a-d10f-49d2-94ae-39c6cd67ea0c" (UID: "ae76828a-d10f-49d2-94ae-39c6cd67ea0c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.192433 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ae76828a-d10f-49d2-94ae-39c6cd67ea0c" (UID: "ae76828a-d10f-49d2-94ae-39c6cd67ea0c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.213614 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.213672 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.213686 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.213701 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b7p5\" (UniqueName: \"kubernetes.io/projected/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-kube-api-access-8b7p5\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.213714 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae76828a-d10f-49d2-94ae-39c6cd67ea0c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.606890 4788 generic.go:334] "Generic (PLEG): container finished" podID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" containerID="227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0" exitCode=0 Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.606943 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" event={"ID":"ae76828a-d10f-49d2-94ae-39c6cd67ea0c","Type":"ContainerDied","Data":"227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0"} Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.607255 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" event={"ID":"ae76828a-d10f-49d2-94ae-39c6cd67ea0c","Type":"ContainerDied","Data":"12650234e6cc244db8fcd26091d15f1836df97ecf91fa61feb7f845e8041035b"} Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.607286 4788 scope.go:117] "RemoveContainer" containerID="227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.606978 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-ccf97d949-s6wcz" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.640216 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-ccf97d949-s6wcz"] Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.649169 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-ccf97d949-s6wcz"] Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.652502 4788 scope.go:117] "RemoveContainer" containerID="75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.671228 4788 scope.go:117] "RemoveContainer" containerID="227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0" Oct 10 16:12:42 crc kubenswrapper[4788]: E1010 16:12:42.671614 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0\": container with ID starting with 227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0 not found: ID does not exist" containerID="227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.671696 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0"} err="failed to get container status \"227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0\": rpc error: code = NotFound desc = could not find container \"227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0\": container with ID starting with 227bb479d0f0c88a6e3e650cbdc99f24b8b7de18cdc0e6ecceae8c30a2bafbd0 not found: ID does not exist" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.671770 4788 scope.go:117] "RemoveContainer" containerID="75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6" Oct 10 16:12:42 crc kubenswrapper[4788]: E1010 16:12:42.672154 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6\": container with ID starting with 75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6 not found: ID does not exist" containerID="75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.672252 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6"} err="failed to get container status \"75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6\": rpc error: code = NotFound desc = could not find container \"75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6\": container with ID starting with 75899aa151ebf455c18308000ec8bb03be60584a7d11499fb0f99180f3e561f6 not found: ID does not exist" Oct 10 16:12:42 crc kubenswrapper[4788]: I1010 16:12:42.906737 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.029400 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-credential-keys\") pod \"8bf92e46-caea-4151-852c-7e29c99bb230\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.029484 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-fernet-keys\") pod \"8bf92e46-caea-4151-852c-7e29c99bb230\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.029548 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-combined-ca-bundle\") pod \"8bf92e46-caea-4151-852c-7e29c99bb230\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.029567 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-config-data\") pod \"8bf92e46-caea-4151-852c-7e29c99bb230\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.030247 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bxzq\" (UniqueName: \"kubernetes.io/projected/8bf92e46-caea-4151-852c-7e29c99bb230-kube-api-access-7bxzq\") pod \"8bf92e46-caea-4151-852c-7e29c99bb230\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.030304 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-scripts\") pod \"8bf92e46-caea-4151-852c-7e29c99bb230\" (UID: \"8bf92e46-caea-4151-852c-7e29c99bb230\") " Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.034219 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8bf92e46-caea-4151-852c-7e29c99bb230" (UID: "8bf92e46-caea-4151-852c-7e29c99bb230"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.034277 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8bf92e46-caea-4151-852c-7e29c99bb230" (UID: "8bf92e46-caea-4151-852c-7e29c99bb230"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.036493 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf92e46-caea-4151-852c-7e29c99bb230-kube-api-access-7bxzq" (OuterVolumeSpecName: "kube-api-access-7bxzq") pod "8bf92e46-caea-4151-852c-7e29c99bb230" (UID: "8bf92e46-caea-4151-852c-7e29c99bb230"). InnerVolumeSpecName "kube-api-access-7bxzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.037022 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-scripts" (OuterVolumeSpecName: "scripts") pod "8bf92e46-caea-4151-852c-7e29c99bb230" (UID: "8bf92e46-caea-4151-852c-7e29c99bb230"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.050675 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bf92e46-caea-4151-852c-7e29c99bb230" (UID: "8bf92e46-caea-4151-852c-7e29c99bb230"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.060939 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-config-data" (OuterVolumeSpecName: "config-data") pod "8bf92e46-caea-4151-852c-7e29c99bb230" (UID: "8bf92e46-caea-4151-852c-7e29c99bb230"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.138305 4788 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.138375 4788 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.138390 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.138404 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.138421 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bxzq\" (UniqueName: \"kubernetes.io/projected/8bf92e46-caea-4151-852c-7e29c99bb230-kube-api-access-7bxzq\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.138438 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bf92e46-caea-4151-852c-7e29c99bb230-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.617122 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sb82b" event={"ID":"8bf92e46-caea-4151-852c-7e29c99bb230","Type":"ContainerDied","Data":"e142f8746bf7406b8d897f01f9eb02e4415feb99e4f25e68330ce82109592454"} Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.617463 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e142f8746bf7406b8d897f01f9eb02e4415feb99e4f25e68330ce82109592454" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.617254 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sb82b" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.719229 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7444b6d9d6-vbfgf"] Oct 10 16:12:43 crc kubenswrapper[4788]: E1010 16:12:43.719727 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" containerName="init" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.719738 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" containerName="init" Oct 10 16:12:43 crc kubenswrapper[4788]: E1010 16:12:43.719756 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf92e46-caea-4151-852c-7e29c99bb230" containerName="keystone-bootstrap" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.719763 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf92e46-caea-4151-852c-7e29c99bb230" containerName="keystone-bootstrap" Oct 10 16:12:43 crc kubenswrapper[4788]: E1010 16:12:43.719777 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" containerName="dnsmasq-dns" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.719785 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" containerName="dnsmasq-dns" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.719969 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" containerName="dnsmasq-dns" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.719979 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf92e46-caea-4151-852c-7e29c99bb230" containerName="keystone-bootstrap" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.722287 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.724695 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.724917 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.725052 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-rvj9h" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.725464 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.738320 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7444b6d9d6-vbfgf"] Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.852284 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-fernet-keys\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.852340 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-credential-keys\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.852386 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-config-data\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.852416 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-scripts\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.852442 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2ms5\" (UniqueName: \"kubernetes.io/projected/325746a3-78bd-4fe7-9ff9-6003f15efc55-kube-api-access-v2ms5\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.852487 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-combined-ca-bundle\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.954080 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-fernet-keys\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.954162 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-credential-keys\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.954208 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-config-data\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.954237 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-scripts\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.954262 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2ms5\" (UniqueName: \"kubernetes.io/projected/325746a3-78bd-4fe7-9ff9-6003f15efc55-kube-api-access-v2ms5\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.954311 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-combined-ca-bundle\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.960676 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-scripts\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.960812 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-fernet-keys\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.960864 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-combined-ca-bundle\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.961364 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-credential-keys\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.961801 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/325746a3-78bd-4fe7-9ff9-6003f15efc55-config-data\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:43 crc kubenswrapper[4788]: I1010 16:12:43.972699 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2ms5\" (UniqueName: \"kubernetes.io/projected/325746a3-78bd-4fe7-9ff9-6003f15efc55-kube-api-access-v2ms5\") pod \"keystone-7444b6d9d6-vbfgf\" (UID: \"325746a3-78bd-4fe7-9ff9-6003f15efc55\") " pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:44 crc kubenswrapper[4788]: I1010 16:12:44.038486 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:44 crc kubenswrapper[4788]: I1010 16:12:44.248965 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae76828a-d10f-49d2-94ae-39c6cd67ea0c" path="/var/lib/kubelet/pods/ae76828a-d10f-49d2-94ae-39c6cd67ea0c/volumes" Oct 10 16:12:44 crc kubenswrapper[4788]: I1010 16:12:44.486717 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7444b6d9d6-vbfgf"] Oct 10 16:12:44 crc kubenswrapper[4788]: W1010 16:12:44.497628 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod325746a3_78bd_4fe7_9ff9_6003f15efc55.slice/crio-1b5dedfdb2c9ee213a09ba98a838dd96a5d53628e94c9bb661e663c4f2f4725c WatchSource:0}: Error finding container 1b5dedfdb2c9ee213a09ba98a838dd96a5d53628e94c9bb661e663c4f2f4725c: Status 404 returned error can't find the container with id 1b5dedfdb2c9ee213a09ba98a838dd96a5d53628e94c9bb661e663c4f2f4725c Oct 10 16:12:44 crc kubenswrapper[4788]: I1010 16:12:44.627934 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7444b6d9d6-vbfgf" event={"ID":"325746a3-78bd-4fe7-9ff9-6003f15efc55","Type":"ContainerStarted","Data":"1b5dedfdb2c9ee213a09ba98a838dd96a5d53628e94c9bb661e663c4f2f4725c"} Oct 10 16:12:45 crc kubenswrapper[4788]: I1010 16:12:45.637786 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7444b6d9d6-vbfgf" event={"ID":"325746a3-78bd-4fe7-9ff9-6003f15efc55","Type":"ContainerStarted","Data":"135129b6d123cbefda02d407f088313578321c91a061b5631299419072468a28"} Oct 10 16:12:45 crc kubenswrapper[4788]: I1010 16:12:45.638077 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:12:45 crc kubenswrapper[4788]: I1010 16:12:45.655949 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7444b6d9d6-vbfgf" podStartSLOduration=2.65592836 podStartE2EDuration="2.65592836s" podCreationTimestamp="2025-10-10 16:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:12:45.653915208 +0000 UTC m=+5268.103630796" watchObservedRunningTime="2025-10-10 16:12:45.65592836 +0000 UTC m=+5268.105643938" Oct 10 16:12:45 crc kubenswrapper[4788]: I1010 16:12:45.661731 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:45 crc kubenswrapper[4788]: I1010 16:12:45.662873 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:45 crc kubenswrapper[4788]: I1010 16:12:45.712334 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:46 crc kubenswrapper[4788]: I1010 16:12:46.695165 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:46 crc kubenswrapper[4788]: I1010 16:12:46.928665 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h69gb"] Oct 10 16:12:48 crc kubenswrapper[4788]: I1010 16:12:48.662803 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h69gb" podUID="a27976f0-d788-4750-be52-148b419d7e9c" containerName="registry-server" containerID="cri-o://40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f" gracePeriod=2 Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.218287 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.355457 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rtms\" (UniqueName: \"kubernetes.io/projected/a27976f0-d788-4750-be52-148b419d7e9c-kube-api-access-9rtms\") pod \"a27976f0-d788-4750-be52-148b419d7e9c\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.355983 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-utilities\") pod \"a27976f0-d788-4750-be52-148b419d7e9c\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.356038 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-catalog-content\") pod \"a27976f0-d788-4750-be52-148b419d7e9c\" (UID: \"a27976f0-d788-4750-be52-148b419d7e9c\") " Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.356964 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-utilities" (OuterVolumeSpecName: "utilities") pod "a27976f0-d788-4750-be52-148b419d7e9c" (UID: "a27976f0-d788-4750-be52-148b419d7e9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.361481 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27976f0-d788-4750-be52-148b419d7e9c-kube-api-access-9rtms" (OuterVolumeSpecName: "kube-api-access-9rtms") pod "a27976f0-d788-4750-be52-148b419d7e9c" (UID: "a27976f0-d788-4750-be52-148b419d7e9c"). InnerVolumeSpecName "kube-api-access-9rtms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.445933 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a27976f0-d788-4750-be52-148b419d7e9c" (UID: "a27976f0-d788-4750-be52-148b419d7e9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.458067 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rtms\" (UniqueName: \"kubernetes.io/projected/a27976f0-d788-4750-be52-148b419d7e9c-kube-api-access-9rtms\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.458115 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.458133 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a27976f0-d788-4750-be52-148b419d7e9c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.678951 4788 generic.go:334] "Generic (PLEG): container finished" podID="a27976f0-d788-4750-be52-148b419d7e9c" containerID="40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f" exitCode=0 Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.679011 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69gb" event={"ID":"a27976f0-d788-4750-be52-148b419d7e9c","Type":"ContainerDied","Data":"40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f"} Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.679023 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h69gb" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.679045 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69gb" event={"ID":"a27976f0-d788-4750-be52-148b419d7e9c","Type":"ContainerDied","Data":"aeb1d9ace0d8fb92e3c4edf459bea53ef7637d977bb5492c63d8facc24e41d8d"} Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.679068 4788 scope.go:117] "RemoveContainer" containerID="40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.707576 4788 scope.go:117] "RemoveContainer" containerID="a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.726782 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h69gb"] Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.740562 4788 scope.go:117] "RemoveContainer" containerID="1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.744087 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h69gb"] Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.793182 4788 scope.go:117] "RemoveContainer" containerID="40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f" Oct 10 16:12:49 crc kubenswrapper[4788]: E1010 16:12:49.793756 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f\": container with ID starting with 40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f not found: ID does not exist" containerID="40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.793914 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f"} err="failed to get container status \"40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f\": rpc error: code = NotFound desc = could not find container \"40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f\": container with ID starting with 40768617863b55975422511f1a244be7efbefdc93974ec69150690f01fefb48f not found: ID does not exist" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.794028 4788 scope.go:117] "RemoveContainer" containerID="a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb" Oct 10 16:12:49 crc kubenswrapper[4788]: E1010 16:12:49.794552 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb\": container with ID starting with a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb not found: ID does not exist" containerID="a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.794588 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb"} err="failed to get container status \"a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb\": rpc error: code = NotFound desc = could not find container \"a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb\": container with ID starting with a326c1134e408db1500405508bf14d975224f97e99e6a1fcf5bc02ac359cbbeb not found: ID does not exist" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.794618 4788 scope.go:117] "RemoveContainer" containerID="1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9" Oct 10 16:12:49 crc kubenswrapper[4788]: E1010 16:12:49.794968 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9\": container with ID starting with 1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9 not found: ID does not exist" containerID="1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9" Oct 10 16:12:49 crc kubenswrapper[4788]: I1010 16:12:49.795099 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9"} err="failed to get container status \"1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9\": rpc error: code = NotFound desc = could not find container \"1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9\": container with ID starting with 1ca3b2ed896866e415c53cc641785432c19be0b03b1d70c43359b4e0e0c4e4f9 not found: ID does not exist" Oct 10 16:12:50 crc kubenswrapper[4788]: I1010 16:12:50.261976 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a27976f0-d788-4750-be52-148b419d7e9c" path="/var/lib/kubelet/pods/a27976f0-d788-4750-be52-148b419d7e9c/volumes" Oct 10 16:13:15 crc kubenswrapper[4788]: I1010 16:13:15.583409 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7444b6d9d6-vbfgf" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.899969 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 10 16:13:19 crc kubenswrapper[4788]: E1010 16:13:19.900702 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27976f0-d788-4750-be52-148b419d7e9c" containerName="extract-content" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.900722 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27976f0-d788-4750-be52-148b419d7e9c" containerName="extract-content" Oct 10 16:13:19 crc kubenswrapper[4788]: E1010 16:13:19.900761 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27976f0-d788-4750-be52-148b419d7e9c" containerName="registry-server" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.900769 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27976f0-d788-4750-be52-148b419d7e9c" containerName="registry-server" Oct 10 16:13:19 crc kubenswrapper[4788]: E1010 16:13:19.900796 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27976f0-d788-4750-be52-148b419d7e9c" containerName="extract-utilities" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.900807 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27976f0-d788-4750-be52-148b419d7e9c" containerName="extract-utilities" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.901021 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27976f0-d788-4750-be52-148b419d7e9c" containerName="registry-server" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.901768 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.905576 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-6l5z4" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.906177 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.907373 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 10 16:13:19 crc kubenswrapper[4788]: I1010 16:13:19.915945 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.030891 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wjjd\" (UniqueName: \"kubernetes.io/projected/7a53dce9-770e-4fbf-8635-f69cc25fafd6-kube-api-access-8wjjd\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.031010 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config-secret\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.031065 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.132877 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wjjd\" (UniqueName: \"kubernetes.io/projected/7a53dce9-770e-4fbf-8635-f69cc25fafd6-kube-api-access-8wjjd\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.132952 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config-secret\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.132978 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.133737 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.148061 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config-secret\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.151294 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wjjd\" (UniqueName: \"kubernetes.io/projected/7a53dce9-770e-4fbf-8635-f69cc25fafd6-kube-api-access-8wjjd\") pod \"openstackclient\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.232537 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 16:13:20 crc kubenswrapper[4788]: I1010 16:13:20.662646 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 16:13:21 crc kubenswrapper[4788]: I1010 16:13:21.010905 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"7a53dce9-770e-4fbf-8635-f69cc25fafd6","Type":"ContainerStarted","Data":"d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c"} Oct 10 16:13:21 crc kubenswrapper[4788]: I1010 16:13:21.011362 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"7a53dce9-770e-4fbf-8635-f69cc25fafd6","Type":"ContainerStarted","Data":"f13ae89c64224e7fecb676a30e72c037ed573d68f48e2b697587995bff030070"} Oct 10 16:13:21 crc kubenswrapper[4788]: I1010 16:13:21.027918 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.027895217 podStartE2EDuration="2.027895217s" podCreationTimestamp="2025-10-10 16:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:13:21.0268455 +0000 UTC m=+5303.476561068" watchObservedRunningTime="2025-10-10 16:13:21.027895217 +0000 UTC m=+5303.477610765" Oct 10 16:14:22 crc kubenswrapper[4788]: I1010 16:14:22.969846 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zfs8j"] Oct 10 16:14:22 crc kubenswrapper[4788]: I1010 16:14:22.973299 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:22 crc kubenswrapper[4788]: I1010 16:14:22.975122 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfs8j"] Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.091561 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-utilities\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.091925 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnkg8\" (UniqueName: \"kubernetes.io/projected/13bb69fe-5afb-47e0-9111-df08bb998dce-kube-api-access-pnkg8\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.092085 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-catalog-content\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.196317 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-catalog-content\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.196394 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-utilities\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.196510 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnkg8\" (UniqueName: \"kubernetes.io/projected/13bb69fe-5afb-47e0-9111-df08bb998dce-kube-api-access-pnkg8\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.197071 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-catalog-content\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.197443 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-utilities\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.218114 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnkg8\" (UniqueName: \"kubernetes.io/projected/13bb69fe-5afb-47e0-9111-df08bb998dce-kube-api-access-pnkg8\") pod \"redhat-marketplace-zfs8j\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.293502 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:23 crc kubenswrapper[4788]: I1010 16:14:23.740671 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfs8j"] Oct 10 16:14:24 crc kubenswrapper[4788]: I1010 16:14:24.600667 4788 generic.go:334] "Generic (PLEG): container finished" podID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerID="6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063" exitCode=0 Oct 10 16:14:24 crc kubenswrapper[4788]: I1010 16:14:24.600754 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfs8j" event={"ID":"13bb69fe-5afb-47e0-9111-df08bb998dce","Type":"ContainerDied","Data":"6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063"} Oct 10 16:14:24 crc kubenswrapper[4788]: I1010 16:14:24.601032 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfs8j" event={"ID":"13bb69fe-5afb-47e0-9111-df08bb998dce","Type":"ContainerStarted","Data":"8feeedd664f4c6d6619755d64e6be82342036c0c08d279e45aedba8529f73d10"} Oct 10 16:14:25 crc kubenswrapper[4788]: I1010 16:14:25.611812 4788 generic.go:334] "Generic (PLEG): container finished" podID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerID="11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb" exitCode=0 Oct 10 16:14:25 crc kubenswrapper[4788]: I1010 16:14:25.612013 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfs8j" event={"ID":"13bb69fe-5afb-47e0-9111-df08bb998dce","Type":"ContainerDied","Data":"11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb"} Oct 10 16:14:27 crc kubenswrapper[4788]: I1010 16:14:27.633590 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfs8j" event={"ID":"13bb69fe-5afb-47e0-9111-df08bb998dce","Type":"ContainerStarted","Data":"5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e"} Oct 10 16:14:27 crc kubenswrapper[4788]: I1010 16:14:27.667103 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zfs8j" podStartSLOduration=3.715607998 podStartE2EDuration="5.667087131s" podCreationTimestamp="2025-10-10 16:14:22 +0000 UTC" firstStartedPulling="2025-10-10 16:14:24.603303034 +0000 UTC m=+5367.053018582" lastFinishedPulling="2025-10-10 16:14:26.554782147 +0000 UTC m=+5369.004497715" observedRunningTime="2025-10-10 16:14:27.660928503 +0000 UTC m=+5370.110644051" watchObservedRunningTime="2025-10-10 16:14:27.667087131 +0000 UTC m=+5370.116802679" Oct 10 16:14:29 crc kubenswrapper[4788]: I1010 16:14:29.406560 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:14:29 crc kubenswrapper[4788]: I1010 16:14:29.407010 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:14:33 crc kubenswrapper[4788]: I1010 16:14:33.294517 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:33 crc kubenswrapper[4788]: I1010 16:14:33.295263 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:33 crc kubenswrapper[4788]: I1010 16:14:33.353850 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:33 crc kubenswrapper[4788]: I1010 16:14:33.738792 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:33 crc kubenswrapper[4788]: I1010 16:14:33.791090 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfs8j"] Oct 10 16:14:35 crc kubenswrapper[4788]: I1010 16:14:35.716702 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zfs8j" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerName="registry-server" containerID="cri-o://5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e" gracePeriod=2 Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.261220 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.352172 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-utilities\") pod \"13bb69fe-5afb-47e0-9111-df08bb998dce\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.352318 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-catalog-content\") pod \"13bb69fe-5afb-47e0-9111-df08bb998dce\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.352529 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnkg8\" (UniqueName: \"kubernetes.io/projected/13bb69fe-5afb-47e0-9111-df08bb998dce-kube-api-access-pnkg8\") pod \"13bb69fe-5afb-47e0-9111-df08bb998dce\" (UID: \"13bb69fe-5afb-47e0-9111-df08bb998dce\") " Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.354429 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-utilities" (OuterVolumeSpecName: "utilities") pod "13bb69fe-5afb-47e0-9111-df08bb998dce" (UID: "13bb69fe-5afb-47e0-9111-df08bb998dce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.366796 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13bb69fe-5afb-47e0-9111-df08bb998dce-kube-api-access-pnkg8" (OuterVolumeSpecName: "kube-api-access-pnkg8") pod "13bb69fe-5afb-47e0-9111-df08bb998dce" (UID: "13bb69fe-5afb-47e0-9111-df08bb998dce"). InnerVolumeSpecName "kube-api-access-pnkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.376185 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13bb69fe-5afb-47e0-9111-df08bb998dce" (UID: "13bb69fe-5afb-47e0-9111-df08bb998dce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.454979 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnkg8\" (UniqueName: \"kubernetes.io/projected/13bb69fe-5afb-47e0-9111-df08bb998dce-kube-api-access-pnkg8\") on node \"crc\" DevicePath \"\"" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.455027 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.455043 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13bb69fe-5afb-47e0-9111-df08bb998dce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.802405 4788 generic.go:334] "Generic (PLEG): container finished" podID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerID="5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e" exitCode=0 Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.802843 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfs8j" event={"ID":"13bb69fe-5afb-47e0-9111-df08bb998dce","Type":"ContainerDied","Data":"5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e"} Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.802885 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zfs8j" event={"ID":"13bb69fe-5afb-47e0-9111-df08bb998dce","Type":"ContainerDied","Data":"8feeedd664f4c6d6619755d64e6be82342036c0c08d279e45aedba8529f73d10"} Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.802908 4788 scope.go:117] "RemoveContainer" containerID="5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.803180 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zfs8j" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.904227 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfs8j"] Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.913148 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zfs8j"] Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.917125 4788 scope.go:117] "RemoveContainer" containerID="11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb" Oct 10 16:14:36 crc kubenswrapper[4788]: I1010 16:14:36.971373 4788 scope.go:117] "RemoveContainer" containerID="6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063" Oct 10 16:14:37 crc kubenswrapper[4788]: I1010 16:14:37.000393 4788 scope.go:117] "RemoveContainer" containerID="5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e" Oct 10 16:14:37 crc kubenswrapper[4788]: E1010 16:14:37.001389 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e\": container with ID starting with 5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e not found: ID does not exist" containerID="5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e" Oct 10 16:14:37 crc kubenswrapper[4788]: I1010 16:14:37.001466 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e"} err="failed to get container status \"5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e\": rpc error: code = NotFound desc = could not find container \"5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e\": container with ID starting with 5e7246400c1f3a5774f3deb8a3cc88b359a12a3c0f88401263b64ea2a3a9ba4e not found: ID does not exist" Oct 10 16:14:37 crc kubenswrapper[4788]: I1010 16:14:37.001512 4788 scope.go:117] "RemoveContainer" containerID="11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb" Oct 10 16:14:37 crc kubenswrapper[4788]: E1010 16:14:37.001901 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb\": container with ID starting with 11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb not found: ID does not exist" containerID="11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb" Oct 10 16:14:37 crc kubenswrapper[4788]: I1010 16:14:37.001970 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb"} err="failed to get container status \"11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb\": rpc error: code = NotFound desc = could not find container \"11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb\": container with ID starting with 11c738bce87d72d95346ac4097d3bbbf9c5369312c1fdc39a01090f224c0a6eb not found: ID does not exist" Oct 10 16:14:37 crc kubenswrapper[4788]: I1010 16:14:37.002018 4788 scope.go:117] "RemoveContainer" containerID="6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063" Oct 10 16:14:37 crc kubenswrapper[4788]: E1010 16:14:37.002483 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063\": container with ID starting with 6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063 not found: ID does not exist" containerID="6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063" Oct 10 16:14:37 crc kubenswrapper[4788]: I1010 16:14:37.002523 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063"} err="failed to get container status \"6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063\": rpc error: code = NotFound desc = could not find container \"6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063\": container with ID starting with 6d8cf02d1108ea70cec84b496eca11c51b7b89747e073dbad83f0b3b1a710063 not found: ID does not exist" Oct 10 16:14:38 crc kubenswrapper[4788]: I1010 16:14:38.245632 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" path="/var/lib/kubelet/pods/13bb69fe-5afb-47e0-9111-df08bb998dce/volumes" Oct 10 16:14:59 crc kubenswrapper[4788]: I1010 16:14:59.405651 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:14:59 crc kubenswrapper[4788]: I1010 16:14:59.406294 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.153231 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff"] Oct 10 16:15:00 crc kubenswrapper[4788]: E1010 16:15:00.154063 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerName="extract-content" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.154083 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerName="extract-content" Oct 10 16:15:00 crc kubenswrapper[4788]: E1010 16:15:00.154114 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerName="extract-utilities" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.154121 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerName="extract-utilities" Oct 10 16:15:00 crc kubenswrapper[4788]: E1010 16:15:00.154133 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerName="registry-server" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.154155 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerName="registry-server" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.154389 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="13bb69fe-5afb-47e0-9111-df08bb998dce" containerName="registry-server" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.155187 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.158235 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.158475 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.161329 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff"] Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.308049 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t92pn\" (UniqueName: \"kubernetes.io/projected/5b54db3b-37c4-45b5-a0ab-b753c63244d2-kube-api-access-t92pn\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.308171 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b54db3b-37c4-45b5-a0ab-b753c63244d2-secret-volume\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.308211 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b54db3b-37c4-45b5-a0ab-b753c63244d2-config-volume\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.410174 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t92pn\" (UniqueName: \"kubernetes.io/projected/5b54db3b-37c4-45b5-a0ab-b753c63244d2-kube-api-access-t92pn\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.410242 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b54db3b-37c4-45b5-a0ab-b753c63244d2-secret-volume\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.410295 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b54db3b-37c4-45b5-a0ab-b753c63244d2-config-volume\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.413459 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b54db3b-37c4-45b5-a0ab-b753c63244d2-config-volume\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.426453 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b54db3b-37c4-45b5-a0ab-b753c63244d2-secret-volume\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.429624 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t92pn\" (UniqueName: \"kubernetes.io/projected/5b54db3b-37c4-45b5-a0ab-b753c63244d2-kube-api-access-t92pn\") pod \"collect-profiles-29335215-vdkff\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.486772 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:00 crc kubenswrapper[4788]: I1010 16:15:00.936297 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff"] Oct 10 16:15:01 crc kubenswrapper[4788]: I1010 16:15:01.065783 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" event={"ID":"5b54db3b-37c4-45b5-a0ab-b753c63244d2","Type":"ContainerStarted","Data":"44ac2a6dff7f48c904480c810e2478e9b91cf8fc82b66fed1257703c3dffc312"} Oct 10 16:15:02 crc kubenswrapper[4788]: I1010 16:15:02.078081 4788 generic.go:334] "Generic (PLEG): container finished" podID="5b54db3b-37c4-45b5-a0ab-b753c63244d2" containerID="d055b2a89ac62efd456f6596a77a1db560101832dead2e6f6300d201c22ab46d" exitCode=0 Oct 10 16:15:02 crc kubenswrapper[4788]: I1010 16:15:02.078216 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" event={"ID":"5b54db3b-37c4-45b5-a0ab-b753c63244d2","Type":"ContainerDied","Data":"d055b2a89ac62efd456f6596a77a1db560101832dead2e6f6300d201c22ab46d"} Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.457814 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.562717 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t92pn\" (UniqueName: \"kubernetes.io/projected/5b54db3b-37c4-45b5-a0ab-b753c63244d2-kube-api-access-t92pn\") pod \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.562796 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b54db3b-37c4-45b5-a0ab-b753c63244d2-config-volume\") pod \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.563032 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b54db3b-37c4-45b5-a0ab-b753c63244d2-secret-volume\") pod \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\" (UID: \"5b54db3b-37c4-45b5-a0ab-b753c63244d2\") " Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.564041 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b54db3b-37c4-45b5-a0ab-b753c63244d2-config-volume" (OuterVolumeSpecName: "config-volume") pod "5b54db3b-37c4-45b5-a0ab-b753c63244d2" (UID: "5b54db3b-37c4-45b5-a0ab-b753c63244d2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.568993 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b54db3b-37c4-45b5-a0ab-b753c63244d2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5b54db3b-37c4-45b5-a0ab-b753c63244d2" (UID: "5b54db3b-37c4-45b5-a0ab-b753c63244d2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.573540 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b54db3b-37c4-45b5-a0ab-b753c63244d2-kube-api-access-t92pn" (OuterVolumeSpecName: "kube-api-access-t92pn") pod "5b54db3b-37c4-45b5-a0ab-b753c63244d2" (UID: "5b54db3b-37c4-45b5-a0ab-b753c63244d2"). InnerVolumeSpecName "kube-api-access-t92pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.665642 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b54db3b-37c4-45b5-a0ab-b753c63244d2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.665726 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t92pn\" (UniqueName: \"kubernetes.io/projected/5b54db3b-37c4-45b5-a0ab-b753c63244d2-kube-api-access-t92pn\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:03 crc kubenswrapper[4788]: I1010 16:15:03.665750 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b54db3b-37c4-45b5-a0ab-b753c63244d2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:04 crc kubenswrapper[4788]: I1010 16:15:04.099979 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" event={"ID":"5b54db3b-37c4-45b5-a0ab-b753c63244d2","Type":"ContainerDied","Data":"44ac2a6dff7f48c904480c810e2478e9b91cf8fc82b66fed1257703c3dffc312"} Oct 10 16:15:04 crc kubenswrapper[4788]: I1010 16:15:04.100022 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44ac2a6dff7f48c904480c810e2478e9b91cf8fc82b66fed1257703c3dffc312" Oct 10 16:15:04 crc kubenswrapper[4788]: I1010 16:15:04.100028 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff" Oct 10 16:15:04 crc kubenswrapper[4788]: I1010 16:15:04.534615 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn"] Oct 10 16:15:04 crc kubenswrapper[4788]: I1010 16:15:04.541036 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335170-px4wn"] Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.666790 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-cmgzc"] Oct 10 16:15:05 crc kubenswrapper[4788]: E1010 16:15:05.667445 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54db3b-37c4-45b5-a0ab-b753c63244d2" containerName="collect-profiles" Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.667458 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54db3b-37c4-45b5-a0ab-b753c63244d2" containerName="collect-profiles" Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.667642 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b54db3b-37c4-45b5-a0ab-b753c63244d2" containerName="collect-profiles" Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.668286 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cmgzc" Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.677245 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-cmgzc"] Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.804544 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6ds5\" (UniqueName: \"kubernetes.io/projected/6890f9fa-fe41-4306-b929-333908049b5b-kube-api-access-m6ds5\") pod \"barbican-db-create-cmgzc\" (UID: \"6890f9fa-fe41-4306-b929-333908049b5b\") " pod="openstack/barbican-db-create-cmgzc" Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.906349 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6ds5\" (UniqueName: \"kubernetes.io/projected/6890f9fa-fe41-4306-b929-333908049b5b-kube-api-access-m6ds5\") pod \"barbican-db-create-cmgzc\" (UID: \"6890f9fa-fe41-4306-b929-333908049b5b\") " pod="openstack/barbican-db-create-cmgzc" Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.929790 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6ds5\" (UniqueName: \"kubernetes.io/projected/6890f9fa-fe41-4306-b929-333908049b5b-kube-api-access-m6ds5\") pod \"barbican-db-create-cmgzc\" (UID: \"6890f9fa-fe41-4306-b929-333908049b5b\") " pod="openstack/barbican-db-create-cmgzc" Oct 10 16:15:05 crc kubenswrapper[4788]: I1010 16:15:05.984434 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cmgzc" Oct 10 16:15:06 crc kubenswrapper[4788]: I1010 16:15:06.243727 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aab10930-60df-4195-b105-da165a8b4fea" path="/var/lib/kubelet/pods/aab10930-60df-4195-b105-da165a8b4fea/volumes" Oct 10 16:15:06 crc kubenswrapper[4788]: I1010 16:15:06.404266 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-cmgzc"] Oct 10 16:15:07 crc kubenswrapper[4788]: I1010 16:15:07.134884 4788 generic.go:334] "Generic (PLEG): container finished" podID="6890f9fa-fe41-4306-b929-333908049b5b" containerID="0a9db373359db6e784c743683b81485d53adfe54e92be33fec9f3c29bf31f207" exitCode=0 Oct 10 16:15:07 crc kubenswrapper[4788]: I1010 16:15:07.134956 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cmgzc" event={"ID":"6890f9fa-fe41-4306-b929-333908049b5b","Type":"ContainerDied","Data":"0a9db373359db6e784c743683b81485d53adfe54e92be33fec9f3c29bf31f207"} Oct 10 16:15:07 crc kubenswrapper[4788]: I1010 16:15:07.135171 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cmgzc" event={"ID":"6890f9fa-fe41-4306-b929-333908049b5b","Type":"ContainerStarted","Data":"b99bd00028eddacd21b399ee393cff53830fab3a717ad0685fce2f733be0b537"} Oct 10 16:15:08 crc kubenswrapper[4788]: I1010 16:15:08.461482 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cmgzc" Oct 10 16:15:08 crc kubenswrapper[4788]: I1010 16:15:08.557132 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6ds5\" (UniqueName: \"kubernetes.io/projected/6890f9fa-fe41-4306-b929-333908049b5b-kube-api-access-m6ds5\") pod \"6890f9fa-fe41-4306-b929-333908049b5b\" (UID: \"6890f9fa-fe41-4306-b929-333908049b5b\") " Oct 10 16:15:08 crc kubenswrapper[4788]: I1010 16:15:08.562472 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6890f9fa-fe41-4306-b929-333908049b5b-kube-api-access-m6ds5" (OuterVolumeSpecName: "kube-api-access-m6ds5") pod "6890f9fa-fe41-4306-b929-333908049b5b" (UID: "6890f9fa-fe41-4306-b929-333908049b5b"). InnerVolumeSpecName "kube-api-access-m6ds5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:15:08 crc kubenswrapper[4788]: I1010 16:15:08.659435 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6ds5\" (UniqueName: \"kubernetes.io/projected/6890f9fa-fe41-4306-b929-333908049b5b-kube-api-access-m6ds5\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:09 crc kubenswrapper[4788]: I1010 16:15:09.153582 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cmgzc" event={"ID":"6890f9fa-fe41-4306-b929-333908049b5b","Type":"ContainerDied","Data":"b99bd00028eddacd21b399ee393cff53830fab3a717ad0685fce2f733be0b537"} Oct 10 16:15:09 crc kubenswrapper[4788]: I1010 16:15:09.153624 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b99bd00028eddacd21b399ee393cff53830fab3a717ad0685fce2f733be0b537" Oct 10 16:15:09 crc kubenswrapper[4788]: I1010 16:15:09.153675 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cmgzc" Oct 10 16:15:12 crc kubenswrapper[4788]: I1010 16:15:12.562128 4788 scope.go:117] "RemoveContainer" containerID="9b000249408e6ca31216e6a3d236e835b3baf6337604721730d3c7be9c80637f" Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.696665 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-75a0-account-create-87fb5"] Oct 10 16:15:15 crc kubenswrapper[4788]: E1010 16:15:15.697645 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6890f9fa-fe41-4306-b929-333908049b5b" containerName="mariadb-database-create" Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.697666 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6890f9fa-fe41-4306-b929-333908049b5b" containerName="mariadb-database-create" Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.697895 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6890f9fa-fe41-4306-b929-333908049b5b" containerName="mariadb-database-create" Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.698599 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-75a0-account-create-87fb5" Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.701970 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.703010 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx4d5\" (UniqueName: \"kubernetes.io/projected/9abf38c2-4b60-49d2-9170-1e94f8f85dcb-kube-api-access-qx4d5\") pod \"barbican-75a0-account-create-87fb5\" (UID: \"9abf38c2-4b60-49d2-9170-1e94f8f85dcb\") " pod="openstack/barbican-75a0-account-create-87fb5" Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.722108 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-75a0-account-create-87fb5"] Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.805716 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx4d5\" (UniqueName: \"kubernetes.io/projected/9abf38c2-4b60-49d2-9170-1e94f8f85dcb-kube-api-access-qx4d5\") pod \"barbican-75a0-account-create-87fb5\" (UID: \"9abf38c2-4b60-49d2-9170-1e94f8f85dcb\") " pod="openstack/barbican-75a0-account-create-87fb5" Oct 10 16:15:15 crc kubenswrapper[4788]: I1010 16:15:15.828888 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx4d5\" (UniqueName: \"kubernetes.io/projected/9abf38c2-4b60-49d2-9170-1e94f8f85dcb-kube-api-access-qx4d5\") pod \"barbican-75a0-account-create-87fb5\" (UID: \"9abf38c2-4b60-49d2-9170-1e94f8f85dcb\") " pod="openstack/barbican-75a0-account-create-87fb5" Oct 10 16:15:16 crc kubenswrapper[4788]: I1010 16:15:16.068612 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-75a0-account-create-87fb5" Oct 10 16:15:16 crc kubenswrapper[4788]: I1010 16:15:16.543469 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-75a0-account-create-87fb5"] Oct 10 16:15:17 crc kubenswrapper[4788]: I1010 16:15:17.236831 4788 generic.go:334] "Generic (PLEG): container finished" podID="9abf38c2-4b60-49d2-9170-1e94f8f85dcb" containerID="cc3c3bd0ded678436d6110125c23a7051d8e416dde9d6e9417129781ad9f2763" exitCode=0 Oct 10 16:15:17 crc kubenswrapper[4788]: I1010 16:15:17.236906 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-75a0-account-create-87fb5" event={"ID":"9abf38c2-4b60-49d2-9170-1e94f8f85dcb","Type":"ContainerDied","Data":"cc3c3bd0ded678436d6110125c23a7051d8e416dde9d6e9417129781ad9f2763"} Oct 10 16:15:17 crc kubenswrapper[4788]: I1010 16:15:17.236945 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-75a0-account-create-87fb5" event={"ID":"9abf38c2-4b60-49d2-9170-1e94f8f85dcb","Type":"ContainerStarted","Data":"28745f79876e7400ddf8c6594fbc42d4444a58984b904a42c0f0d67bbec137b7"} Oct 10 16:15:18 crc kubenswrapper[4788]: I1010 16:15:18.609794 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-75a0-account-create-87fb5" Oct 10 16:15:18 crc kubenswrapper[4788]: I1010 16:15:18.765628 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx4d5\" (UniqueName: \"kubernetes.io/projected/9abf38c2-4b60-49d2-9170-1e94f8f85dcb-kube-api-access-qx4d5\") pod \"9abf38c2-4b60-49d2-9170-1e94f8f85dcb\" (UID: \"9abf38c2-4b60-49d2-9170-1e94f8f85dcb\") " Oct 10 16:15:18 crc kubenswrapper[4788]: I1010 16:15:18.771569 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9abf38c2-4b60-49d2-9170-1e94f8f85dcb-kube-api-access-qx4d5" (OuterVolumeSpecName: "kube-api-access-qx4d5") pod "9abf38c2-4b60-49d2-9170-1e94f8f85dcb" (UID: "9abf38c2-4b60-49d2-9170-1e94f8f85dcb"). InnerVolumeSpecName "kube-api-access-qx4d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:15:18 crc kubenswrapper[4788]: I1010 16:15:18.867679 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx4d5\" (UniqueName: \"kubernetes.io/projected/9abf38c2-4b60-49d2-9170-1e94f8f85dcb-kube-api-access-qx4d5\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:19 crc kubenswrapper[4788]: I1010 16:15:19.260966 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-75a0-account-create-87fb5" event={"ID":"9abf38c2-4b60-49d2-9170-1e94f8f85dcb","Type":"ContainerDied","Data":"28745f79876e7400ddf8c6594fbc42d4444a58984b904a42c0f0d67bbec137b7"} Oct 10 16:15:19 crc kubenswrapper[4788]: I1010 16:15:19.261025 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-75a0-account-create-87fb5" Oct 10 16:15:19 crc kubenswrapper[4788]: I1010 16:15:19.261026 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28745f79876e7400ddf8c6594fbc42d4444a58984b904a42c0f0d67bbec137b7" Oct 10 16:15:20 crc kubenswrapper[4788]: I1010 16:15:20.925320 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-z6wgq"] Oct 10 16:15:20 crc kubenswrapper[4788]: E1010 16:15:20.926656 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9abf38c2-4b60-49d2-9170-1e94f8f85dcb" containerName="mariadb-account-create" Oct 10 16:15:20 crc kubenswrapper[4788]: I1010 16:15:20.926682 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9abf38c2-4b60-49d2-9170-1e94f8f85dcb" containerName="mariadb-account-create" Oct 10 16:15:20 crc kubenswrapper[4788]: I1010 16:15:20.926996 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9abf38c2-4b60-49d2-9170-1e94f8f85dcb" containerName="mariadb-account-create" Oct 10 16:15:20 crc kubenswrapper[4788]: I1010 16:15:20.928003 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:20 crc kubenswrapper[4788]: I1010 16:15:20.930844 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-q8cmt" Oct 10 16:15:20 crc kubenswrapper[4788]: I1010 16:15:20.931034 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 10 16:15:20 crc kubenswrapper[4788]: I1010 16:15:20.943395 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-z6wgq"] Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.007028 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-combined-ca-bundle\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.007087 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4tk7\" (UniqueName: \"kubernetes.io/projected/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-kube-api-access-b4tk7\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.007159 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-db-sync-config-data\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.109089 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-combined-ca-bundle\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.109155 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4tk7\" (UniqueName: \"kubernetes.io/projected/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-kube-api-access-b4tk7\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.109214 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-db-sync-config-data\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.116770 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-db-sync-config-data\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.119202 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-combined-ca-bundle\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.143656 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4tk7\" (UniqueName: \"kubernetes.io/projected/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-kube-api-access-b4tk7\") pod \"barbican-db-sync-z6wgq\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.264909 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:21 crc kubenswrapper[4788]: I1010 16:15:21.733065 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-z6wgq"] Oct 10 16:15:22 crc kubenswrapper[4788]: I1010 16:15:22.295075 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z6wgq" event={"ID":"e2189f91-f8ed-47ef-8b6e-00b20b6b996b","Type":"ContainerStarted","Data":"055df83035bd2bd6fd838d33bfe726f548d044dafe230f490f21fb5bf0cd3d83"} Oct 10 16:15:22 crc kubenswrapper[4788]: I1010 16:15:22.295469 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z6wgq" event={"ID":"e2189f91-f8ed-47ef-8b6e-00b20b6b996b","Type":"ContainerStarted","Data":"0d9ac3c4313d389c64672f1d328ee1b4fd23bb43a50ef98074fa50b3597124be"} Oct 10 16:15:22 crc kubenswrapper[4788]: I1010 16:15:22.318863 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-z6wgq" podStartSLOduration=2.318840081 podStartE2EDuration="2.318840081s" podCreationTimestamp="2025-10-10 16:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:15:22.310561058 +0000 UTC m=+5424.760276606" watchObservedRunningTime="2025-10-10 16:15:22.318840081 +0000 UTC m=+5424.768555639" Oct 10 16:15:23 crc kubenswrapper[4788]: I1010 16:15:23.306779 4788 generic.go:334] "Generic (PLEG): container finished" podID="e2189f91-f8ed-47ef-8b6e-00b20b6b996b" containerID="055df83035bd2bd6fd838d33bfe726f548d044dafe230f490f21fb5bf0cd3d83" exitCode=0 Oct 10 16:15:23 crc kubenswrapper[4788]: I1010 16:15:23.306874 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z6wgq" event={"ID":"e2189f91-f8ed-47ef-8b6e-00b20b6b996b","Type":"ContainerDied","Data":"055df83035bd2bd6fd838d33bfe726f548d044dafe230f490f21fb5bf0cd3d83"} Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.700432 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.780092 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-combined-ca-bundle\") pod \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.780242 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-db-sync-config-data\") pod \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.780280 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4tk7\" (UniqueName: \"kubernetes.io/projected/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-kube-api-access-b4tk7\") pod \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\" (UID: \"e2189f91-f8ed-47ef-8b6e-00b20b6b996b\") " Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.785677 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-kube-api-access-b4tk7" (OuterVolumeSpecName: "kube-api-access-b4tk7") pod "e2189f91-f8ed-47ef-8b6e-00b20b6b996b" (UID: "e2189f91-f8ed-47ef-8b6e-00b20b6b996b"). InnerVolumeSpecName "kube-api-access-b4tk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.785775 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e2189f91-f8ed-47ef-8b6e-00b20b6b996b" (UID: "e2189f91-f8ed-47ef-8b6e-00b20b6b996b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.801999 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2189f91-f8ed-47ef-8b6e-00b20b6b996b" (UID: "e2189f91-f8ed-47ef-8b6e-00b20b6b996b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.882210 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.882241 4788 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:24 crc kubenswrapper[4788]: I1010 16:15:24.882250 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4tk7\" (UniqueName: \"kubernetes.io/projected/e2189f91-f8ed-47ef-8b6e-00b20b6b996b-kube-api-access-b4tk7\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.331700 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-z6wgq" event={"ID":"e2189f91-f8ed-47ef-8b6e-00b20b6b996b","Type":"ContainerDied","Data":"0d9ac3c4313d389c64672f1d328ee1b4fd23bb43a50ef98074fa50b3597124be"} Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.331761 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d9ac3c4313d389c64672f1d328ee1b4fd23bb43a50ef98074fa50b3597124be" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.331842 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-z6wgq" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.550825 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-86b75f4667-rhtqc"] Oct 10 16:15:25 crc kubenswrapper[4788]: E1010 16:15:25.551394 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2189f91-f8ed-47ef-8b6e-00b20b6b996b" containerName="barbican-db-sync" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.551416 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2189f91-f8ed-47ef-8b6e-00b20b6b996b" containerName="barbican-db-sync" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.551681 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2189f91-f8ed-47ef-8b6e-00b20b6b996b" containerName="barbican-db-sync" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.558578 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.561810 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.563733 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.564032 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-q8cmt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.583070 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-86b75f4667-rhtqc"] Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.610012 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-656647f4cd-wp2dt"] Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.612053 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.621850 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.632795 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-656647f4cd-wp2dt"] Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.695678 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56fb8b755-5l2mm"] Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.698243 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.701079 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-config-data-custom\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.710298 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbf57fa7-9537-46e6-9902-ea78b098412e-logs\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.710582 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-combined-ca-bundle\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.710705 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-config-data-custom\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.710861 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-config-data\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.710961 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-combined-ca-bundle\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.711075 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-config-data\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.711184 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stdzc\" (UniqueName: \"kubernetes.io/projected/bbf57fa7-9537-46e6-9902-ea78b098412e-kube-api-access-stdzc\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.711347 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bmrs\" (UniqueName: \"kubernetes.io/projected/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-kube-api-access-6bmrs\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.711502 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-logs\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.717980 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56fb8b755-5l2mm"] Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814081 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-config-data\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814126 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stdzc\" (UniqueName: \"kubernetes.io/projected/bbf57fa7-9537-46e6-9902-ea78b098412e-kube-api-access-stdzc\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814207 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bmrs\" (UniqueName: \"kubernetes.io/projected/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-kube-api-access-6bmrs\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814230 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-logs\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814252 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j2xx\" (UniqueName: \"kubernetes.io/projected/4fcafd24-9d93-47db-985a-103c354e8f2b-kube-api-access-5j2xx\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814272 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-sb\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814294 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-dns-svc\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814328 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-config-data-custom\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814344 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbf57fa7-9537-46e6-9902-ea78b098412e-logs\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814380 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-config\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814406 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-combined-ca-bundle\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814424 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-config-data-custom\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814460 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-config-data\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814488 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-combined-ca-bundle\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.814535 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-nb\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.815567 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-logs\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.818599 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbf57fa7-9537-46e6-9902-ea78b098412e-logs\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.833822 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-combined-ca-bundle\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.834505 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-config-data-custom\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.835525 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-config-data\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.835996 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-config-data-custom\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.852103 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8578c7f644-zrr9p"] Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.854765 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bmrs\" (UniqueName: \"kubernetes.io/projected/d8e57d00-cebc-4ea5-9033-c9009ed5c21f-kube-api-access-6bmrs\") pod \"barbican-worker-86b75f4667-rhtqc\" (UID: \"d8e57d00-cebc-4ea5-9033-c9009ed5c21f\") " pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.855212 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-combined-ca-bundle\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.856479 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbf57fa7-9537-46e6-9902-ea78b098412e-config-data\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.870999 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.876040 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stdzc\" (UniqueName: \"kubernetes.io/projected/bbf57fa7-9537-46e6-9902-ea78b098412e-kube-api-access-stdzc\") pod \"barbican-keystone-listener-656647f4cd-wp2dt\" (UID: \"bbf57fa7-9537-46e6-9902-ea78b098412e\") " pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.876693 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-86b75f4667-rhtqc" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.885244 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.916403 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-dns-svc\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.916515 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-config\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.916592 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-nb\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.916669 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j2xx\" (UniqueName: \"kubernetes.io/projected/4fcafd24-9d93-47db-985a-103c354e8f2b-kube-api-access-5j2xx\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.916694 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-sb\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.917987 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-sb\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.918568 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-dns-svc\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.918738 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-nb\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.919380 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-config\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.933221 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8578c7f644-zrr9p"] Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.943783 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" Oct 10 16:15:25 crc kubenswrapper[4788]: I1010 16:15:25.972994 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j2xx\" (UniqueName: \"kubernetes.io/projected/4fcafd24-9d93-47db-985a-103c354e8f2b-kube-api-access-5j2xx\") pod \"dnsmasq-dns-56fb8b755-5l2mm\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.025570 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r74w\" (UniqueName: \"kubernetes.io/projected/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-kube-api-access-8r74w\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.025635 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-config-data-custom\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.025750 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-logs\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.025788 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-config-data\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.025832 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-combined-ca-bundle\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.031580 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.127981 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-logs\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.128031 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-config-data\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.128073 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-combined-ca-bundle\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.128119 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r74w\" (UniqueName: \"kubernetes.io/projected/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-kube-api-access-8r74w\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.128154 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-config-data-custom\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.128878 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-logs\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.133936 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-combined-ca-bundle\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.134703 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-config-data\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.140934 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-config-data-custom\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.153675 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r74w\" (UniqueName: \"kubernetes.io/projected/77ccc4ea-3d0d-46b3-aa0f-ca0affce488c-kube-api-access-8r74w\") pod \"barbican-api-8578c7f644-zrr9p\" (UID: \"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c\") " pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.258118 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.488162 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-86b75f4667-rhtqc"] Oct 10 16:15:26 crc kubenswrapper[4788]: W1010 16:15:26.496229 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8e57d00_cebc_4ea5_9033_c9009ed5c21f.slice/crio-4ea78015e202c50541937c23b8ee488b177dfe793d181673f16a41afc3a61ba4 WatchSource:0}: Error finding container 4ea78015e202c50541937c23b8ee488b177dfe793d181673f16a41afc3a61ba4: Status 404 returned error can't find the container with id 4ea78015e202c50541937c23b8ee488b177dfe793d181673f16a41afc3a61ba4 Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.527830 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8578c7f644-zrr9p"] Oct 10 16:15:26 crc kubenswrapper[4788]: W1010 16:15:26.531371 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ccc4ea_3d0d_46b3_aa0f_ca0affce488c.slice/crio-185df8dcd458e08a4113a3be14980ef7f00528c648e093fa5f45c42201c57748 WatchSource:0}: Error finding container 185df8dcd458e08a4113a3be14980ef7f00528c648e093fa5f45c42201c57748: Status 404 returned error can't find the container with id 185df8dcd458e08a4113a3be14980ef7f00528c648e093fa5f45c42201c57748 Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.586876 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-656647f4cd-wp2dt"] Oct 10 16:15:26 crc kubenswrapper[4788]: I1010 16:15:26.637386 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56fb8b755-5l2mm"] Oct 10 16:15:26 crc kubenswrapper[4788]: W1010 16:15:26.671273 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fcafd24_9d93_47db_985a_103c354e8f2b.slice/crio-2c2cfa1cb1f725d5c040996f6e6922d18d3b3429d01696f9312eb4f85ea0ca16 WatchSource:0}: Error finding container 2c2cfa1cb1f725d5c040996f6e6922d18d3b3429d01696f9312eb4f85ea0ca16: Status 404 returned error can't find the container with id 2c2cfa1cb1f725d5c040996f6e6922d18d3b3429d01696f9312eb4f85ea0ca16 Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.376107 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8578c7f644-zrr9p" event={"ID":"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c","Type":"ContainerStarted","Data":"d47832bd865d2197b63311a92f58f2d9fcf8ccea47f77908e6a02af6e4789662"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.376204 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8578c7f644-zrr9p" event={"ID":"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c","Type":"ContainerStarted","Data":"a50b9a2b5444fe9c41bd99ca2553cbe057cab6573f14a424ad190fa29ec7c2bc"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.376220 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8578c7f644-zrr9p" event={"ID":"77ccc4ea-3d0d-46b3-aa0f-ca0affce488c","Type":"ContainerStarted","Data":"185df8dcd458e08a4113a3be14980ef7f00528c648e093fa5f45c42201c57748"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.378289 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.378337 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.379489 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" event={"ID":"bbf57fa7-9537-46e6-9902-ea78b098412e","Type":"ContainerStarted","Data":"e87a864fce1154e4d10c5478fec28bb022e96efe9be0959b0240888318978571"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.379546 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" event={"ID":"bbf57fa7-9537-46e6-9902-ea78b098412e","Type":"ContainerStarted","Data":"d4de6c319016c5b71ef7b3efcb6e5ea694a51d8117187b6f89d6a0468867c2f5"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.379563 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" event={"ID":"bbf57fa7-9537-46e6-9902-ea78b098412e","Type":"ContainerStarted","Data":"d588bd33ea03553a5201a54c269d944c3c8f6c4714aa93400600d3ed88d53a7b"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.382487 4788 generic.go:334] "Generic (PLEG): container finished" podID="4fcafd24-9d93-47db-985a-103c354e8f2b" containerID="0f08c629f86d5d95d5a1f957bd8fc72877a2f7effcb67bab9688b8aad15ef652" exitCode=0 Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.382581 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" event={"ID":"4fcafd24-9d93-47db-985a-103c354e8f2b","Type":"ContainerDied","Data":"0f08c629f86d5d95d5a1f957bd8fc72877a2f7effcb67bab9688b8aad15ef652"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.382643 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" event={"ID":"4fcafd24-9d93-47db-985a-103c354e8f2b","Type":"ContainerStarted","Data":"2c2cfa1cb1f725d5c040996f6e6922d18d3b3429d01696f9312eb4f85ea0ca16"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.410418 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-86b75f4667-rhtqc" event={"ID":"d8e57d00-cebc-4ea5-9033-c9009ed5c21f","Type":"ContainerStarted","Data":"e02af1a44b49676ffbbcd576199d75e619b2e7184ae8c572d02f6fbdfeac0d85"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.410473 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-86b75f4667-rhtqc" event={"ID":"d8e57d00-cebc-4ea5-9033-c9009ed5c21f","Type":"ContainerStarted","Data":"1291c16674c8a48f5531f6226f52cc9a71163c3c523e79ff6ff0dc0bb525b463"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.410482 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-86b75f4667-rhtqc" event={"ID":"d8e57d00-cebc-4ea5-9033-c9009ed5c21f","Type":"ContainerStarted","Data":"4ea78015e202c50541937c23b8ee488b177dfe793d181673f16a41afc3a61ba4"} Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.428182 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8578c7f644-zrr9p" podStartSLOduration=2.4281145459999998 podStartE2EDuration="2.428114546s" podCreationTimestamp="2025-10-10 16:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:15:27.411242953 +0000 UTC m=+5429.860958521" watchObservedRunningTime="2025-10-10 16:15:27.428114546 +0000 UTC m=+5429.877830094" Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.499073 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-86b75f4667-rhtqc" podStartSLOduration=2.499041869 podStartE2EDuration="2.499041869s" podCreationTimestamp="2025-10-10 16:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:15:27.480912883 +0000 UTC m=+5429.930628431" watchObservedRunningTime="2025-10-10 16:15:27.499041869 +0000 UTC m=+5429.948757417" Oct 10 16:15:27 crc kubenswrapper[4788]: I1010 16:15:27.505932 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-656647f4cd-wp2dt" podStartSLOduration=2.505901485 podStartE2EDuration="2.505901485s" podCreationTimestamp="2025-10-10 16:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:15:27.502693423 +0000 UTC m=+5429.952408981" watchObservedRunningTime="2025-10-10 16:15:27.505901485 +0000 UTC m=+5429.955617033" Oct 10 16:15:28 crc kubenswrapper[4788]: I1010 16:15:28.420105 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" event={"ID":"4fcafd24-9d93-47db-985a-103c354e8f2b","Type":"ContainerStarted","Data":"3248e4ad59f02f6307414ace17cb898d5a02790437e08e0bd7aa1285c0188d0d"} Oct 10 16:15:28 crc kubenswrapper[4788]: I1010 16:15:28.421216 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:28 crc kubenswrapper[4788]: I1010 16:15:28.437907 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" podStartSLOduration=3.437889448 podStartE2EDuration="3.437889448s" podCreationTimestamp="2025-10-10 16:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:15:28.435400594 +0000 UTC m=+5430.885116142" watchObservedRunningTime="2025-10-10 16:15:28.437889448 +0000 UTC m=+5430.887604996" Oct 10 16:15:29 crc kubenswrapper[4788]: I1010 16:15:29.406694 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:15:29 crc kubenswrapper[4788]: I1010 16:15:29.406774 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:15:29 crc kubenswrapper[4788]: I1010 16:15:29.406836 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:15:29 crc kubenswrapper[4788]: I1010 16:15:29.407845 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2a936fbf95226f44505aa4c234e52e71830372cdc6c6f705ee31bf2d04b44c1"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:15:29 crc kubenswrapper[4788]: I1010 16:15:29.407946 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://f2a936fbf95226f44505aa4c234e52e71830372cdc6c6f705ee31bf2d04b44c1" gracePeriod=600 Oct 10 16:15:30 crc kubenswrapper[4788]: I1010 16:15:30.439409 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="f2a936fbf95226f44505aa4c234e52e71830372cdc6c6f705ee31bf2d04b44c1" exitCode=0 Oct 10 16:15:30 crc kubenswrapper[4788]: I1010 16:15:30.439447 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"f2a936fbf95226f44505aa4c234e52e71830372cdc6c6f705ee31bf2d04b44c1"} Oct 10 16:15:30 crc kubenswrapper[4788]: I1010 16:15:30.440172 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2"} Oct 10 16:15:30 crc kubenswrapper[4788]: I1010 16:15:30.440200 4788 scope.go:117] "RemoveContainer" containerID="9d87f4c2e45583fdca4868f5931cb4e1b62fbd53ac973ab2aa6f7c33e8e08e9e" Oct 10 16:15:32 crc kubenswrapper[4788]: I1010 16:15:32.751271 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:34 crc kubenswrapper[4788]: I1010 16:15:34.131637 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8578c7f644-zrr9p" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.033603 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.085629 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99458d6fc-ngrpb"] Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.086472 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" podUID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerName="dnsmasq-dns" containerID="cri-o://91022de13ed6443c1cd593c5d52d8ac0f587968ce49cc5d9078bb3eaeb741f40" gracePeriod=10 Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.515350 4788 generic.go:334] "Generic (PLEG): container finished" podID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerID="91022de13ed6443c1cd593c5d52d8ac0f587968ce49cc5d9078bb3eaeb741f40" exitCode=0 Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.515722 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" event={"ID":"ef832418-6747-45e3-8a6f-4f07eb2b5107","Type":"ContainerDied","Data":"91022de13ed6443c1cd593c5d52d8ac0f587968ce49cc5d9078bb3eaeb741f40"} Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.626920 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.762336 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-sb\") pod \"ef832418-6747-45e3-8a6f-4f07eb2b5107\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.762564 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-config\") pod \"ef832418-6747-45e3-8a6f-4f07eb2b5107\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.762619 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-dns-svc\") pod \"ef832418-6747-45e3-8a6f-4f07eb2b5107\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.762704 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj78w\" (UniqueName: \"kubernetes.io/projected/ef832418-6747-45e3-8a6f-4f07eb2b5107-kube-api-access-zj78w\") pod \"ef832418-6747-45e3-8a6f-4f07eb2b5107\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.762734 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-nb\") pod \"ef832418-6747-45e3-8a6f-4f07eb2b5107\" (UID: \"ef832418-6747-45e3-8a6f-4f07eb2b5107\") " Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.773023 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef832418-6747-45e3-8a6f-4f07eb2b5107-kube-api-access-zj78w" (OuterVolumeSpecName: "kube-api-access-zj78w") pod "ef832418-6747-45e3-8a6f-4f07eb2b5107" (UID: "ef832418-6747-45e3-8a6f-4f07eb2b5107"). InnerVolumeSpecName "kube-api-access-zj78w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.812483 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef832418-6747-45e3-8a6f-4f07eb2b5107" (UID: "ef832418-6747-45e3-8a6f-4f07eb2b5107"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.816097 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef832418-6747-45e3-8a6f-4f07eb2b5107" (UID: "ef832418-6747-45e3-8a6f-4f07eb2b5107"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.818922 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef832418-6747-45e3-8a6f-4f07eb2b5107" (UID: "ef832418-6747-45e3-8a6f-4f07eb2b5107"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.825875 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-config" (OuterVolumeSpecName: "config") pod "ef832418-6747-45e3-8a6f-4f07eb2b5107" (UID: "ef832418-6747-45e3-8a6f-4f07eb2b5107"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.865668 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj78w\" (UniqueName: \"kubernetes.io/projected/ef832418-6747-45e3-8a6f-4f07eb2b5107-kube-api-access-zj78w\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.865786 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.865799 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.865810 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:36 crc kubenswrapper[4788]: I1010 16:15:36.865821 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef832418-6747-45e3-8a6f-4f07eb2b5107-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:37 crc kubenswrapper[4788]: I1010 16:15:37.524468 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" event={"ID":"ef832418-6747-45e3-8a6f-4f07eb2b5107","Type":"ContainerDied","Data":"f6f2290a2170eea648e9a3b32365e920a616996efc4d87ce503b02cd8ee51566"} Oct 10 16:15:37 crc kubenswrapper[4788]: I1010 16:15:37.524559 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" Oct 10 16:15:37 crc kubenswrapper[4788]: I1010 16:15:37.524779 4788 scope.go:117] "RemoveContainer" containerID="91022de13ed6443c1cd593c5d52d8ac0f587968ce49cc5d9078bb3eaeb741f40" Oct 10 16:15:37 crc kubenswrapper[4788]: I1010 16:15:37.549090 4788 scope.go:117] "RemoveContainer" containerID="322a9cdb7fd931e27e56a653ac4d0d77311acc7367d9433d880454a178ec8c01" Oct 10 16:15:37 crc kubenswrapper[4788]: I1010 16:15:37.560099 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99458d6fc-ngrpb"] Oct 10 16:15:37 crc kubenswrapper[4788]: I1010 16:15:37.566502 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-99458d6fc-ngrpb"] Oct 10 16:15:38 crc kubenswrapper[4788]: I1010 16:15:38.252398 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef832418-6747-45e3-8a6f-4f07eb2b5107" path="/var/lib/kubelet/pods/ef832418-6747-45e3-8a6f-4f07eb2b5107/volumes" Oct 10 16:15:41 crc kubenswrapper[4788]: I1010 16:15:41.420638 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-99458d6fc-ngrpb" podUID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.21:5353: i/o timeout" Oct 10 16:15:47 crc kubenswrapper[4788]: I1010 16:15:47.987306 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-m7hfp"] Oct 10 16:15:47 crc kubenswrapper[4788]: E1010 16:15:47.988165 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerName="dnsmasq-dns" Oct 10 16:15:47 crc kubenswrapper[4788]: I1010 16:15:47.988179 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerName="dnsmasq-dns" Oct 10 16:15:47 crc kubenswrapper[4788]: E1010 16:15:47.988194 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerName="init" Oct 10 16:15:47 crc kubenswrapper[4788]: I1010 16:15:47.988213 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerName="init" Oct 10 16:15:47 crc kubenswrapper[4788]: I1010 16:15:47.988399 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef832418-6747-45e3-8a6f-4f07eb2b5107" containerName="dnsmasq-dns" Oct 10 16:15:47 crc kubenswrapper[4788]: I1010 16:15:47.988977 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m7hfp" Oct 10 16:15:48 crc kubenswrapper[4788]: I1010 16:15:47.999965 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-m7hfp"] Oct 10 16:15:48 crc kubenswrapper[4788]: I1010 16:15:48.089124 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9tsv\" (UniqueName: \"kubernetes.io/projected/585409fe-03f3-49b9-959b-6dffa35aa9ed-kube-api-access-x9tsv\") pod \"neutron-db-create-m7hfp\" (UID: \"585409fe-03f3-49b9-959b-6dffa35aa9ed\") " pod="openstack/neutron-db-create-m7hfp" Oct 10 16:15:48 crc kubenswrapper[4788]: I1010 16:15:48.191322 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9tsv\" (UniqueName: \"kubernetes.io/projected/585409fe-03f3-49b9-959b-6dffa35aa9ed-kube-api-access-x9tsv\") pod \"neutron-db-create-m7hfp\" (UID: \"585409fe-03f3-49b9-959b-6dffa35aa9ed\") " pod="openstack/neutron-db-create-m7hfp" Oct 10 16:15:48 crc kubenswrapper[4788]: I1010 16:15:48.210330 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9tsv\" (UniqueName: \"kubernetes.io/projected/585409fe-03f3-49b9-959b-6dffa35aa9ed-kube-api-access-x9tsv\") pod \"neutron-db-create-m7hfp\" (UID: \"585409fe-03f3-49b9-959b-6dffa35aa9ed\") " pod="openstack/neutron-db-create-m7hfp" Oct 10 16:15:48 crc kubenswrapper[4788]: I1010 16:15:48.317004 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m7hfp" Oct 10 16:15:48 crc kubenswrapper[4788]: I1010 16:15:48.769991 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-m7hfp"] Oct 10 16:15:49 crc kubenswrapper[4788]: I1010 16:15:49.652262 4788 generic.go:334] "Generic (PLEG): container finished" podID="585409fe-03f3-49b9-959b-6dffa35aa9ed" containerID="23ed7bc67039c1344f91e2d4882f93ad0bb327feed0c8911b6809f25be4bca9f" exitCode=0 Oct 10 16:15:49 crc kubenswrapper[4788]: I1010 16:15:49.652361 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m7hfp" event={"ID":"585409fe-03f3-49b9-959b-6dffa35aa9ed","Type":"ContainerDied","Data":"23ed7bc67039c1344f91e2d4882f93ad0bb327feed0c8911b6809f25be4bca9f"} Oct 10 16:15:49 crc kubenswrapper[4788]: I1010 16:15:49.652627 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m7hfp" event={"ID":"585409fe-03f3-49b9-959b-6dffa35aa9ed","Type":"ContainerStarted","Data":"c86cc6860c11616fe997ea23ceddde13a714d64cc92e098f5316547ec102ab09"} Oct 10 16:15:50 crc kubenswrapper[4788]: I1010 16:15:50.980913 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m7hfp" Oct 10 16:15:51 crc kubenswrapper[4788]: I1010 16:15:51.140665 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9tsv\" (UniqueName: \"kubernetes.io/projected/585409fe-03f3-49b9-959b-6dffa35aa9ed-kube-api-access-x9tsv\") pod \"585409fe-03f3-49b9-959b-6dffa35aa9ed\" (UID: \"585409fe-03f3-49b9-959b-6dffa35aa9ed\") " Oct 10 16:15:51 crc kubenswrapper[4788]: I1010 16:15:51.148944 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/585409fe-03f3-49b9-959b-6dffa35aa9ed-kube-api-access-x9tsv" (OuterVolumeSpecName: "kube-api-access-x9tsv") pod "585409fe-03f3-49b9-959b-6dffa35aa9ed" (UID: "585409fe-03f3-49b9-959b-6dffa35aa9ed"). InnerVolumeSpecName "kube-api-access-x9tsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:15:51 crc kubenswrapper[4788]: I1010 16:15:51.242360 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9tsv\" (UniqueName: \"kubernetes.io/projected/585409fe-03f3-49b9-959b-6dffa35aa9ed-kube-api-access-x9tsv\") on node \"crc\" DevicePath \"\"" Oct 10 16:15:51 crc kubenswrapper[4788]: I1010 16:15:51.671776 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m7hfp" event={"ID":"585409fe-03f3-49b9-959b-6dffa35aa9ed","Type":"ContainerDied","Data":"c86cc6860c11616fe997ea23ceddde13a714d64cc92e098f5316547ec102ab09"} Oct 10 16:15:51 crc kubenswrapper[4788]: I1010 16:15:51.672075 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c86cc6860c11616fe997ea23ceddde13a714d64cc92e098f5316547ec102ab09" Oct 10 16:15:51 crc kubenswrapper[4788]: I1010 16:15:51.671830 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m7hfp" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.095720 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6858-account-create-h5vvl"] Oct 10 16:15:58 crc kubenswrapper[4788]: E1010 16:15:58.096387 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="585409fe-03f3-49b9-959b-6dffa35aa9ed" containerName="mariadb-database-create" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.096400 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="585409fe-03f3-49b9-959b-6dffa35aa9ed" containerName="mariadb-database-create" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.096562 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="585409fe-03f3-49b9-959b-6dffa35aa9ed" containerName="mariadb-database-create" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.097200 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6858-account-create-h5vvl" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.101576 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.108566 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6858-account-create-h5vvl"] Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.189965 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6d58\" (UniqueName: \"kubernetes.io/projected/d7264913-5014-4307-a0eb-cd635b68a152-kube-api-access-s6d58\") pod \"neutron-6858-account-create-h5vvl\" (UID: \"d7264913-5014-4307-a0eb-cd635b68a152\") " pod="openstack/neutron-6858-account-create-h5vvl" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.291482 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6d58\" (UniqueName: \"kubernetes.io/projected/d7264913-5014-4307-a0eb-cd635b68a152-kube-api-access-s6d58\") pod \"neutron-6858-account-create-h5vvl\" (UID: \"d7264913-5014-4307-a0eb-cd635b68a152\") " pod="openstack/neutron-6858-account-create-h5vvl" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.317738 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6d58\" (UniqueName: \"kubernetes.io/projected/d7264913-5014-4307-a0eb-cd635b68a152-kube-api-access-s6d58\") pod \"neutron-6858-account-create-h5vvl\" (UID: \"d7264913-5014-4307-a0eb-cd635b68a152\") " pod="openstack/neutron-6858-account-create-h5vvl" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.469236 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6858-account-create-h5vvl" Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.890200 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6858-account-create-h5vvl"] Oct 10 16:15:58 crc kubenswrapper[4788]: I1010 16:15:58.900230 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 10 16:15:59 crc kubenswrapper[4788]: I1010 16:15:59.753023 4788 generic.go:334] "Generic (PLEG): container finished" podID="d7264913-5014-4307-a0eb-cd635b68a152" containerID="77f1cb29ac193cced6eb4d49370d8b6d596e97131fdf59d267c0d4904374e209" exitCode=0 Oct 10 16:15:59 crc kubenswrapper[4788]: I1010 16:15:59.753062 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6858-account-create-h5vvl" event={"ID":"d7264913-5014-4307-a0eb-cd635b68a152","Type":"ContainerDied","Data":"77f1cb29ac193cced6eb4d49370d8b6d596e97131fdf59d267c0d4904374e209"} Oct 10 16:15:59 crc kubenswrapper[4788]: I1010 16:15:59.753330 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6858-account-create-h5vvl" event={"ID":"d7264913-5014-4307-a0eb-cd635b68a152","Type":"ContainerStarted","Data":"ffea273a6784d408a83308952bc00f9ec3bf652eccb783e53eaafc6253a5b5ac"} Oct 10 16:16:01 crc kubenswrapper[4788]: I1010 16:16:01.052326 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6858-account-create-h5vvl" Oct 10 16:16:01 crc kubenswrapper[4788]: I1010 16:16:01.161076 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6d58\" (UniqueName: \"kubernetes.io/projected/d7264913-5014-4307-a0eb-cd635b68a152-kube-api-access-s6d58\") pod \"d7264913-5014-4307-a0eb-cd635b68a152\" (UID: \"d7264913-5014-4307-a0eb-cd635b68a152\") " Oct 10 16:16:01 crc kubenswrapper[4788]: I1010 16:16:01.166259 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7264913-5014-4307-a0eb-cd635b68a152-kube-api-access-s6d58" (OuterVolumeSpecName: "kube-api-access-s6d58") pod "d7264913-5014-4307-a0eb-cd635b68a152" (UID: "d7264913-5014-4307-a0eb-cd635b68a152"). InnerVolumeSpecName "kube-api-access-s6d58". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:16:01 crc kubenswrapper[4788]: I1010 16:16:01.263268 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6d58\" (UniqueName: \"kubernetes.io/projected/d7264913-5014-4307-a0eb-cd635b68a152-kube-api-access-s6d58\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:01 crc kubenswrapper[4788]: I1010 16:16:01.780548 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6858-account-create-h5vvl" event={"ID":"d7264913-5014-4307-a0eb-cd635b68a152","Type":"ContainerDied","Data":"ffea273a6784d408a83308952bc00f9ec3bf652eccb783e53eaafc6253a5b5ac"} Oct 10 16:16:01 crc kubenswrapper[4788]: I1010 16:16:01.780864 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffea273a6784d408a83308952bc00f9ec3bf652eccb783e53eaafc6253a5b5ac" Oct 10 16:16:01 crc kubenswrapper[4788]: I1010 16:16:01.780617 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6858-account-create-h5vvl" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.331811 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4x65t"] Oct 10 16:16:03 crc kubenswrapper[4788]: E1010 16:16:03.332454 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7264913-5014-4307-a0eb-cd635b68a152" containerName="mariadb-account-create" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.332477 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7264913-5014-4307-a0eb-cd635b68a152" containerName="mariadb-account-create" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.332782 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7264913-5014-4307-a0eb-cd635b68a152" containerName="mariadb-account-create" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.333843 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.337097 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nmkvm" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.337248 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.337449 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.344308 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4x65t"] Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.401972 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-combined-ca-bundle\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.402043 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thvkj\" (UniqueName: \"kubernetes.io/projected/1e118dac-a272-4a53-ac59-1ae9291402aa-kube-api-access-thvkj\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.402153 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-config\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.503716 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thvkj\" (UniqueName: \"kubernetes.io/projected/1e118dac-a272-4a53-ac59-1ae9291402aa-kube-api-access-thvkj\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.503811 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-config\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.503941 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-combined-ca-bundle\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.522392 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-combined-ca-bundle\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.522538 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-config\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.527257 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thvkj\" (UniqueName: \"kubernetes.io/projected/1e118dac-a272-4a53-ac59-1ae9291402aa-kube-api-access-thvkj\") pod \"neutron-db-sync-4x65t\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:03 crc kubenswrapper[4788]: I1010 16:16:03.664736 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:04 crc kubenswrapper[4788]: I1010 16:16:04.109621 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4x65t"] Oct 10 16:16:04 crc kubenswrapper[4788]: I1010 16:16:04.808858 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4x65t" event={"ID":"1e118dac-a272-4a53-ac59-1ae9291402aa","Type":"ContainerStarted","Data":"dc551ed9855cf30e4f6fb3f25b37bfac335dd065e3a0713736c4806f4896ea20"} Oct 10 16:16:04 crc kubenswrapper[4788]: I1010 16:16:04.809443 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4x65t" event={"ID":"1e118dac-a272-4a53-ac59-1ae9291402aa","Type":"ContainerStarted","Data":"9d3207c4c27df9a9c90c4c511355a2724163dbcc40a8259c7e44529e51460c99"} Oct 10 16:16:04 crc kubenswrapper[4788]: I1010 16:16:04.827916 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4x65t" podStartSLOduration=1.827881138 podStartE2EDuration="1.827881138s" podCreationTimestamp="2025-10-10 16:16:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:16:04.826489492 +0000 UTC m=+5467.276205060" watchObservedRunningTime="2025-10-10 16:16:04.827881138 +0000 UTC m=+5467.277596686" Oct 10 16:16:08 crc kubenswrapper[4788]: I1010 16:16:08.847527 4788 generic.go:334] "Generic (PLEG): container finished" podID="1e118dac-a272-4a53-ac59-1ae9291402aa" containerID="dc551ed9855cf30e4f6fb3f25b37bfac335dd065e3a0713736c4806f4896ea20" exitCode=0 Oct 10 16:16:08 crc kubenswrapper[4788]: I1010 16:16:08.847611 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4x65t" event={"ID":"1e118dac-a272-4a53-ac59-1ae9291402aa","Type":"ContainerDied","Data":"dc551ed9855cf30e4f6fb3f25b37bfac335dd065e3a0713736c4806f4896ea20"} Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.179411 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.340060 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-config\") pod \"1e118dac-a272-4a53-ac59-1ae9291402aa\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.340551 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thvkj\" (UniqueName: \"kubernetes.io/projected/1e118dac-a272-4a53-ac59-1ae9291402aa-kube-api-access-thvkj\") pod \"1e118dac-a272-4a53-ac59-1ae9291402aa\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.340656 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-combined-ca-bundle\") pod \"1e118dac-a272-4a53-ac59-1ae9291402aa\" (UID: \"1e118dac-a272-4a53-ac59-1ae9291402aa\") " Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.346567 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e118dac-a272-4a53-ac59-1ae9291402aa-kube-api-access-thvkj" (OuterVolumeSpecName: "kube-api-access-thvkj") pod "1e118dac-a272-4a53-ac59-1ae9291402aa" (UID: "1e118dac-a272-4a53-ac59-1ae9291402aa"). InnerVolumeSpecName "kube-api-access-thvkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.364751 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e118dac-a272-4a53-ac59-1ae9291402aa" (UID: "1e118dac-a272-4a53-ac59-1ae9291402aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.371430 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-config" (OuterVolumeSpecName: "config") pod "1e118dac-a272-4a53-ac59-1ae9291402aa" (UID: "1e118dac-a272-4a53-ac59-1ae9291402aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.442903 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thvkj\" (UniqueName: \"kubernetes.io/projected/1e118dac-a272-4a53-ac59-1ae9291402aa-kube-api-access-thvkj\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.442966 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.442980 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1e118dac-a272-4a53-ac59-1ae9291402aa-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.873842 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4x65t" event={"ID":"1e118dac-a272-4a53-ac59-1ae9291402aa","Type":"ContainerDied","Data":"9d3207c4c27df9a9c90c4c511355a2724163dbcc40a8259c7e44529e51460c99"} Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.873913 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d3207c4c27df9a9c90c4c511355a2724163dbcc40a8259c7e44529e51460c99" Oct 10 16:16:10 crc kubenswrapper[4788]: I1010 16:16:10.874034 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4x65t" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.037954 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d6f69859c-knp8p"] Oct 10 16:16:11 crc kubenswrapper[4788]: E1010 16:16:11.038591 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e118dac-a272-4a53-ac59-1ae9291402aa" containerName="neutron-db-sync" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.038700 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e118dac-a272-4a53-ac59-1ae9291402aa" containerName="neutron-db-sync" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.039023 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e118dac-a272-4a53-ac59-1ae9291402aa" containerName="neutron-db-sync" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.040574 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.053692 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-nb\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.053738 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-sb\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.053768 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-config\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.053822 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-dns-svc\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.053858 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb965\" (UniqueName: \"kubernetes.io/projected/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-kube-api-access-tb965\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.081997 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d6f69859c-knp8p"] Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.154402 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-dns-svc\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.154468 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb965\" (UniqueName: \"kubernetes.io/projected/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-kube-api-access-tb965\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.154518 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-nb\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.154539 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-sb\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.154569 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-config\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.155485 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-config\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.156001 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-dns-svc\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.156764 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-nb\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.157289 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-sb\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.187488 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb965\" (UniqueName: \"kubernetes.io/projected/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-kube-api-access-tb965\") pod \"dnsmasq-dns-d6f69859c-knp8p\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.192564 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-56d899f5c5-sdv5n"] Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.193903 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.198154 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.200220 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.206370 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nmkvm" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.207818 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56d899f5c5-sdv5n"] Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.256215 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-httpd-config\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.256257 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grrzs\" (UniqueName: \"kubernetes.io/projected/4057ec67-9bd6-43d0-941c-39d3373caa74-kube-api-access-grrzs\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.256474 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-config\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.256619 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-combined-ca-bundle\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.358111 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-combined-ca-bundle\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.358244 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-httpd-config\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.358283 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grrzs\" (UniqueName: \"kubernetes.io/projected/4057ec67-9bd6-43d0-941c-39d3373caa74-kube-api-access-grrzs\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.358425 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-config\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.363688 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-httpd-config\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.363707 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-config\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.374320 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.381229 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4057ec67-9bd6-43d0-941c-39d3373caa74-combined-ca-bundle\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.384288 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grrzs\" (UniqueName: \"kubernetes.io/projected/4057ec67-9bd6-43d0-941c-39d3373caa74-kube-api-access-grrzs\") pod \"neutron-56d899f5c5-sdv5n\" (UID: \"4057ec67-9bd6-43d0-941c-39d3373caa74\") " pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.560479 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.828118 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d6f69859c-knp8p"] Oct 10 16:16:11 crc kubenswrapper[4788]: I1010 16:16:11.883685 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" event={"ID":"68cde0e7-40e4-4f5a-a6f3-50d10996a32e","Type":"ContainerStarted","Data":"afde627d7650c7d23a05918a7006b5914815db5fe44ce4890f9284a5456242d0"} Oct 10 16:16:12 crc kubenswrapper[4788]: I1010 16:16:12.169740 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56d899f5c5-sdv5n"] Oct 10 16:16:12 crc kubenswrapper[4788]: I1010 16:16:12.894747 4788 generic.go:334] "Generic (PLEG): container finished" podID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" containerID="73cb1240d9c3acf5bde0bafff031bafedcf5abe442c9ae2f03c6698583940898" exitCode=0 Oct 10 16:16:12 crc kubenswrapper[4788]: I1010 16:16:12.894808 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" event={"ID":"68cde0e7-40e4-4f5a-a6f3-50d10996a32e","Type":"ContainerDied","Data":"73cb1240d9c3acf5bde0bafff031bafedcf5abe442c9ae2f03c6698583940898"} Oct 10 16:16:12 crc kubenswrapper[4788]: I1010 16:16:12.899022 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d899f5c5-sdv5n" event={"ID":"4057ec67-9bd6-43d0-941c-39d3373caa74","Type":"ContainerStarted","Data":"c920230b9991fabb8a836b6f1c5c7993e211fbb88534b6554da694edae75663c"} Oct 10 16:16:12 crc kubenswrapper[4788]: I1010 16:16:12.899071 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d899f5c5-sdv5n" event={"ID":"4057ec67-9bd6-43d0-941c-39d3373caa74","Type":"ContainerStarted","Data":"927921246f9862430e3b741f028b0fb86d2814c98d6cbca20a3ce8aaff50adab"} Oct 10 16:16:12 crc kubenswrapper[4788]: I1010 16:16:12.899088 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56d899f5c5-sdv5n" event={"ID":"4057ec67-9bd6-43d0-941c-39d3373caa74","Type":"ContainerStarted","Data":"526ab76fceeaab189735ca461b32db14a5fefa9660e88863abcefcb852017a4b"} Oct 10 16:16:12 crc kubenswrapper[4788]: I1010 16:16:12.899360 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:12 crc kubenswrapper[4788]: I1010 16:16:12.950310 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-56d899f5c5-sdv5n" podStartSLOduration=1.9502851410000002 podStartE2EDuration="1.950285141s" podCreationTimestamp="2025-10-10 16:16:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:16:12.942381877 +0000 UTC m=+5475.392097425" watchObservedRunningTime="2025-10-10 16:16:12.950285141 +0000 UTC m=+5475.400000709" Oct 10 16:16:13 crc kubenswrapper[4788]: I1010 16:16:13.912262 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" event={"ID":"68cde0e7-40e4-4f5a-a6f3-50d10996a32e","Type":"ContainerStarted","Data":"3232984f0d4d5643f20432f86faa2c36415b2693817ef1b7fa65447d79b458d6"} Oct 10 16:16:13 crc kubenswrapper[4788]: I1010 16:16:13.941211 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" podStartSLOduration=3.941189497 podStartE2EDuration="3.941189497s" podCreationTimestamp="2025-10-10 16:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:16:13.933971841 +0000 UTC m=+5476.383687389" watchObservedRunningTime="2025-10-10 16:16:13.941189497 +0000 UTC m=+5476.390905055" Oct 10 16:16:14 crc kubenswrapper[4788]: I1010 16:16:14.918509 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:21 crc kubenswrapper[4788]: I1010 16:16:21.376431 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:16:21 crc kubenswrapper[4788]: I1010 16:16:21.465966 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56fb8b755-5l2mm"] Oct 10 16:16:21 crc kubenswrapper[4788]: I1010 16:16:21.466834 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" podUID="4fcafd24-9d93-47db-985a-103c354e8f2b" containerName="dnsmasq-dns" containerID="cri-o://3248e4ad59f02f6307414ace17cb898d5a02790437e08e0bd7aa1285c0188d0d" gracePeriod=10 Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:21.999713 4788 generic.go:334] "Generic (PLEG): container finished" podID="4fcafd24-9d93-47db-985a-103c354e8f2b" containerID="3248e4ad59f02f6307414ace17cb898d5a02790437e08e0bd7aa1285c0188d0d" exitCode=0 Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:21.999855 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" event={"ID":"4fcafd24-9d93-47db-985a-103c354e8f2b","Type":"ContainerDied","Data":"3248e4ad59f02f6307414ace17cb898d5a02790437e08e0bd7aa1285c0188d0d"} Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.232786 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.377683 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-nb\") pod \"4fcafd24-9d93-47db-985a-103c354e8f2b\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.377772 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-dns-svc\") pod \"4fcafd24-9d93-47db-985a-103c354e8f2b\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.377892 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-sb\") pod \"4fcafd24-9d93-47db-985a-103c354e8f2b\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.378048 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-config\") pod \"4fcafd24-9d93-47db-985a-103c354e8f2b\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.378100 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j2xx\" (UniqueName: \"kubernetes.io/projected/4fcafd24-9d93-47db-985a-103c354e8f2b-kube-api-access-5j2xx\") pod \"4fcafd24-9d93-47db-985a-103c354e8f2b\" (UID: \"4fcafd24-9d93-47db-985a-103c354e8f2b\") " Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.411625 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fcafd24-9d93-47db-985a-103c354e8f2b-kube-api-access-5j2xx" (OuterVolumeSpecName: "kube-api-access-5j2xx") pod "4fcafd24-9d93-47db-985a-103c354e8f2b" (UID: "4fcafd24-9d93-47db-985a-103c354e8f2b"). InnerVolumeSpecName "kube-api-access-5j2xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.426047 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4fcafd24-9d93-47db-985a-103c354e8f2b" (UID: "4fcafd24-9d93-47db-985a-103c354e8f2b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.429477 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4fcafd24-9d93-47db-985a-103c354e8f2b" (UID: "4fcafd24-9d93-47db-985a-103c354e8f2b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.448037 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-config" (OuterVolumeSpecName: "config") pod "4fcafd24-9d93-47db-985a-103c354e8f2b" (UID: "4fcafd24-9d93-47db-985a-103c354e8f2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.451797 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4fcafd24-9d93-47db-985a-103c354e8f2b" (UID: "4fcafd24-9d93-47db-985a-103c354e8f2b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.480232 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.480275 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.480289 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.480301 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4fcafd24-9d93-47db-985a-103c354e8f2b-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:22 crc kubenswrapper[4788]: I1010 16:16:22.480317 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j2xx\" (UniqueName: \"kubernetes.io/projected/4fcafd24-9d93-47db-985a-103c354e8f2b-kube-api-access-5j2xx\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:23 crc kubenswrapper[4788]: I1010 16:16:23.013940 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" event={"ID":"4fcafd24-9d93-47db-985a-103c354e8f2b","Type":"ContainerDied","Data":"2c2cfa1cb1f725d5c040996f6e6922d18d3b3429d01696f9312eb4f85ea0ca16"} Oct 10 16:16:23 crc kubenswrapper[4788]: I1010 16:16:23.014446 4788 scope.go:117] "RemoveContainer" containerID="3248e4ad59f02f6307414ace17cb898d5a02790437e08e0bd7aa1285c0188d0d" Oct 10 16:16:23 crc kubenswrapper[4788]: I1010 16:16:23.014059 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56fb8b755-5l2mm" Oct 10 16:16:23 crc kubenswrapper[4788]: I1010 16:16:23.056505 4788 scope.go:117] "RemoveContainer" containerID="0f08c629f86d5d95d5a1f957bd8fc72877a2f7effcb67bab9688b8aad15ef652" Oct 10 16:16:23 crc kubenswrapper[4788]: I1010 16:16:23.074513 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56fb8b755-5l2mm"] Oct 10 16:16:23 crc kubenswrapper[4788]: I1010 16:16:23.085323 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56fb8b755-5l2mm"] Oct 10 16:16:24 crc kubenswrapper[4788]: I1010 16:16:24.247078 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fcafd24-9d93-47db-985a-103c354e8f2b" path="/var/lib/kubelet/pods/4fcafd24-9d93-47db-985a-103c354e8f2b/volumes" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.161859 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-clmxg"] Oct 10 16:16:31 crc kubenswrapper[4788]: E1010 16:16:31.163080 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fcafd24-9d93-47db-985a-103c354e8f2b" containerName="init" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.163096 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fcafd24-9d93-47db-985a-103c354e8f2b" containerName="init" Oct 10 16:16:31 crc kubenswrapper[4788]: E1010 16:16:31.163157 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fcafd24-9d93-47db-985a-103c354e8f2b" containerName="dnsmasq-dns" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.163163 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fcafd24-9d93-47db-985a-103c354e8f2b" containerName="dnsmasq-dns" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.163344 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fcafd24-9d93-47db-985a-103c354e8f2b" containerName="dnsmasq-dns" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.164716 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.181555 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clmxg"] Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.262198 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-catalog-content\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.262574 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t7nt\" (UniqueName: \"kubernetes.io/projected/8239a070-d3f5-4059-ac54-2b6c5b8d899e-kube-api-access-6t7nt\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.262859 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-utilities\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.364888 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-utilities\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.365019 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-catalog-content\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.365102 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t7nt\" (UniqueName: \"kubernetes.io/projected/8239a070-d3f5-4059-ac54-2b6c5b8d899e-kube-api-access-6t7nt\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.365605 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-utilities\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.365711 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-catalog-content\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.391612 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t7nt\" (UniqueName: \"kubernetes.io/projected/8239a070-d3f5-4059-ac54-2b6c5b8d899e-kube-api-access-6t7nt\") pod \"community-operators-clmxg\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:31 crc kubenswrapper[4788]: I1010 16:16:31.501982 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:32 crc kubenswrapper[4788]: I1010 16:16:32.098398 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clmxg"] Oct 10 16:16:33 crc kubenswrapper[4788]: I1010 16:16:33.113909 4788 generic.go:334] "Generic (PLEG): container finished" podID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerID="412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48" exitCode=0 Oct 10 16:16:33 crc kubenswrapper[4788]: I1010 16:16:33.114318 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clmxg" event={"ID":"8239a070-d3f5-4059-ac54-2b6c5b8d899e","Type":"ContainerDied","Data":"412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48"} Oct 10 16:16:33 crc kubenswrapper[4788]: I1010 16:16:33.114389 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clmxg" event={"ID":"8239a070-d3f5-4059-ac54-2b6c5b8d899e","Type":"ContainerStarted","Data":"64e43ff852ddf9345a3437bd689079f2bf5dcaba6da2ba182c129d1aa0973386"} Oct 10 16:16:37 crc kubenswrapper[4788]: I1010 16:16:37.152221 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clmxg" event={"ID":"8239a070-d3f5-4059-ac54-2b6c5b8d899e","Type":"ContainerStarted","Data":"5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909"} Oct 10 16:16:38 crc kubenswrapper[4788]: I1010 16:16:38.161108 4788 generic.go:334] "Generic (PLEG): container finished" podID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerID="5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909" exitCode=0 Oct 10 16:16:38 crc kubenswrapper[4788]: I1010 16:16:38.161190 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clmxg" event={"ID":"8239a070-d3f5-4059-ac54-2b6c5b8d899e","Type":"ContainerDied","Data":"5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909"} Oct 10 16:16:39 crc kubenswrapper[4788]: I1010 16:16:39.176818 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clmxg" event={"ID":"8239a070-d3f5-4059-ac54-2b6c5b8d899e","Type":"ContainerStarted","Data":"bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608"} Oct 10 16:16:39 crc kubenswrapper[4788]: I1010 16:16:39.210673 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-clmxg" podStartSLOduration=2.777883836 podStartE2EDuration="8.210633212s" podCreationTimestamp="2025-10-10 16:16:31 +0000 UTC" firstStartedPulling="2025-10-10 16:16:33.117219464 +0000 UTC m=+5495.566935012" lastFinishedPulling="2025-10-10 16:16:38.54996884 +0000 UTC m=+5500.999684388" observedRunningTime="2025-10-10 16:16:39.206779414 +0000 UTC m=+5501.656494992" watchObservedRunningTime="2025-10-10 16:16:39.210633212 +0000 UTC m=+5501.660348810" Oct 10 16:16:41 crc kubenswrapper[4788]: I1010 16:16:41.502672 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:41 crc kubenswrapper[4788]: I1010 16:16:41.503093 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:41 crc kubenswrapper[4788]: I1010 16:16:41.555436 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:41 crc kubenswrapper[4788]: I1010 16:16:41.578591 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-56d899f5c5-sdv5n" Oct 10 16:16:49 crc kubenswrapper[4788]: I1010 16:16:49.819688 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6ndzr"] Oct 10 16:16:49 crc kubenswrapper[4788]: I1010 16:16:49.822060 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6ndzr" Oct 10 16:16:49 crc kubenswrapper[4788]: I1010 16:16:49.828684 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6ndzr"] Oct 10 16:16:49 crc kubenswrapper[4788]: I1010 16:16:49.957923 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bqv6\" (UniqueName: \"kubernetes.io/projected/e755581d-b00e-41d1-bb31-e4384bbf444f-kube-api-access-7bqv6\") pod \"glance-db-create-6ndzr\" (UID: \"e755581d-b00e-41d1-bb31-e4384bbf444f\") " pod="openstack/glance-db-create-6ndzr" Oct 10 16:16:50 crc kubenswrapper[4788]: I1010 16:16:50.060030 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bqv6\" (UniqueName: \"kubernetes.io/projected/e755581d-b00e-41d1-bb31-e4384bbf444f-kube-api-access-7bqv6\") pod \"glance-db-create-6ndzr\" (UID: \"e755581d-b00e-41d1-bb31-e4384bbf444f\") " pod="openstack/glance-db-create-6ndzr" Oct 10 16:16:50 crc kubenswrapper[4788]: I1010 16:16:50.088004 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bqv6\" (UniqueName: \"kubernetes.io/projected/e755581d-b00e-41d1-bb31-e4384bbf444f-kube-api-access-7bqv6\") pod \"glance-db-create-6ndzr\" (UID: \"e755581d-b00e-41d1-bb31-e4384bbf444f\") " pod="openstack/glance-db-create-6ndzr" Oct 10 16:16:50 crc kubenswrapper[4788]: I1010 16:16:50.144487 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6ndzr" Oct 10 16:16:50 crc kubenswrapper[4788]: I1010 16:16:50.602237 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6ndzr"] Oct 10 16:16:50 crc kubenswrapper[4788]: W1010 16:16:50.604344 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode755581d_b00e_41d1_bb31_e4384bbf444f.slice/crio-79b46c6d36ba846813586b6a6e7231fc55f14fb689e405652dc47a7226d5d393 WatchSource:0}: Error finding container 79b46c6d36ba846813586b6a6e7231fc55f14fb689e405652dc47a7226d5d393: Status 404 returned error can't find the container with id 79b46c6d36ba846813586b6a6e7231fc55f14fb689e405652dc47a7226d5d393 Oct 10 16:16:51 crc kubenswrapper[4788]: I1010 16:16:51.322482 4788 generic.go:334] "Generic (PLEG): container finished" podID="e755581d-b00e-41d1-bb31-e4384bbf444f" containerID="f4331ca8c2fa2441de223d36d991407d7f1588ed8124c6e47353b5868391dfc7" exitCode=0 Oct 10 16:16:51 crc kubenswrapper[4788]: I1010 16:16:51.322587 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6ndzr" event={"ID":"e755581d-b00e-41d1-bb31-e4384bbf444f","Type":"ContainerDied","Data":"f4331ca8c2fa2441de223d36d991407d7f1588ed8124c6e47353b5868391dfc7"} Oct 10 16:16:51 crc kubenswrapper[4788]: I1010 16:16:51.322868 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6ndzr" event={"ID":"e755581d-b00e-41d1-bb31-e4384bbf444f","Type":"ContainerStarted","Data":"79b46c6d36ba846813586b6a6e7231fc55f14fb689e405652dc47a7226d5d393"} Oct 10 16:16:51 crc kubenswrapper[4788]: I1010 16:16:51.561522 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-clmxg" Oct 10 16:16:51 crc kubenswrapper[4788]: I1010 16:16:51.661829 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clmxg"] Oct 10 16:16:51 crc kubenswrapper[4788]: I1010 16:16:51.701680 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fmcvh"] Oct 10 16:16:51 crc kubenswrapper[4788]: I1010 16:16:51.702004 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fmcvh" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerName="registry-server" containerID="cri-o://34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803" gracePeriod=2 Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.167595 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmcvh" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.214739 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-catalog-content\") pod \"c86646c0-727e-4893-9d74-888fd96d2fa7\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.214786 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g78z\" (UniqueName: \"kubernetes.io/projected/c86646c0-727e-4893-9d74-888fd96d2fa7-kube-api-access-9g78z\") pod \"c86646c0-727e-4893-9d74-888fd96d2fa7\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.215027 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-utilities\") pod \"c86646c0-727e-4893-9d74-888fd96d2fa7\" (UID: \"c86646c0-727e-4893-9d74-888fd96d2fa7\") " Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.216895 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-utilities" (OuterVolumeSpecName: "utilities") pod "c86646c0-727e-4893-9d74-888fd96d2fa7" (UID: "c86646c0-727e-4893-9d74-888fd96d2fa7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.224354 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c86646c0-727e-4893-9d74-888fd96d2fa7-kube-api-access-9g78z" (OuterVolumeSpecName: "kube-api-access-9g78z") pod "c86646c0-727e-4893-9d74-888fd96d2fa7" (UID: "c86646c0-727e-4893-9d74-888fd96d2fa7"). InnerVolumeSpecName "kube-api-access-9g78z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.303613 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c86646c0-727e-4893-9d74-888fd96d2fa7" (UID: "c86646c0-727e-4893-9d74-888fd96d2fa7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.317033 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.317066 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g78z\" (UniqueName: \"kubernetes.io/projected/c86646c0-727e-4893-9d74-888fd96d2fa7-kube-api-access-9g78z\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.317080 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86646c0-727e-4893-9d74-888fd96d2fa7-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.350570 4788 generic.go:334] "Generic (PLEG): container finished" podID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerID="34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803" exitCode=0 Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.351736 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmcvh" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.353282 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcvh" event={"ID":"c86646c0-727e-4893-9d74-888fd96d2fa7","Type":"ContainerDied","Data":"34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803"} Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.353769 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmcvh" event={"ID":"c86646c0-727e-4893-9d74-888fd96d2fa7","Type":"ContainerDied","Data":"7b9220823dc8130a1b8cee697c7cae8433198fa0c714500e540881b7e8156d27"} Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.353796 4788 scope.go:117] "RemoveContainer" containerID="34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.377347 4788 scope.go:117] "RemoveContainer" containerID="acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.414663 4788 scope.go:117] "RemoveContainer" containerID="d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.414743 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fmcvh"] Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.415234 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fmcvh"] Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.454518 4788 scope.go:117] "RemoveContainer" containerID="34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803" Oct 10 16:16:52 crc kubenswrapper[4788]: E1010 16:16:52.455100 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803\": container with ID starting with 34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803 not found: ID does not exist" containerID="34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.455164 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803"} err="failed to get container status \"34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803\": rpc error: code = NotFound desc = could not find container \"34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803\": container with ID starting with 34d5a23e3b47fe75227bab50a0e15a7d3f265870d234ec21891939ca8989f803 not found: ID does not exist" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.455199 4788 scope.go:117] "RemoveContainer" containerID="acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9" Oct 10 16:16:52 crc kubenswrapper[4788]: E1010 16:16:52.455528 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9\": container with ID starting with acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9 not found: ID does not exist" containerID="acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.455550 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9"} err="failed to get container status \"acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9\": rpc error: code = NotFound desc = could not find container \"acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9\": container with ID starting with acc899c6aeed93caaf0e03e980092b257f0dd0d6eda5aa38c78c873a14c600c9 not found: ID does not exist" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.455563 4788 scope.go:117] "RemoveContainer" containerID="d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8" Oct 10 16:16:52 crc kubenswrapper[4788]: E1010 16:16:52.455787 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8\": container with ID starting with d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8 not found: ID does not exist" containerID="d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.455807 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8"} err="failed to get container status \"d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8\": rpc error: code = NotFound desc = could not find container \"d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8\": container with ID starting with d83caacaffa564b8fb054c8720a8b49d307683db273aa79708851d95917058e8 not found: ID does not exist" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.677716 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6ndzr" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.727522 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bqv6\" (UniqueName: \"kubernetes.io/projected/e755581d-b00e-41d1-bb31-e4384bbf444f-kube-api-access-7bqv6\") pod \"e755581d-b00e-41d1-bb31-e4384bbf444f\" (UID: \"e755581d-b00e-41d1-bb31-e4384bbf444f\") " Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.738649 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e755581d-b00e-41d1-bb31-e4384bbf444f-kube-api-access-7bqv6" (OuterVolumeSpecName: "kube-api-access-7bqv6") pod "e755581d-b00e-41d1-bb31-e4384bbf444f" (UID: "e755581d-b00e-41d1-bb31-e4384bbf444f"). InnerVolumeSpecName "kube-api-access-7bqv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:16:52 crc kubenswrapper[4788]: I1010 16:16:52.830710 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bqv6\" (UniqueName: \"kubernetes.io/projected/e755581d-b00e-41d1-bb31-e4384bbf444f-kube-api-access-7bqv6\") on node \"crc\" DevicePath \"\"" Oct 10 16:16:53 crc kubenswrapper[4788]: I1010 16:16:53.363411 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6ndzr" event={"ID":"e755581d-b00e-41d1-bb31-e4384bbf444f","Type":"ContainerDied","Data":"79b46c6d36ba846813586b6a6e7231fc55f14fb689e405652dc47a7226d5d393"} Oct 10 16:16:53 crc kubenswrapper[4788]: I1010 16:16:53.363472 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79b46c6d36ba846813586b6a6e7231fc55f14fb689e405652dc47a7226d5d393" Oct 10 16:16:53 crc kubenswrapper[4788]: I1010 16:16:53.363482 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6ndzr" Oct 10 16:16:54 crc kubenswrapper[4788]: I1010 16:16:54.245164 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" path="/var/lib/kubelet/pods/c86646c0-727e-4893-9d74-888fd96d2fa7/volumes" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.910043 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-9279-account-create-6hfhf"] Oct 10 16:16:59 crc kubenswrapper[4788]: E1010 16:16:59.910902 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e755581d-b00e-41d1-bb31-e4384bbf444f" containerName="mariadb-database-create" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.910927 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e755581d-b00e-41d1-bb31-e4384bbf444f" containerName="mariadb-database-create" Oct 10 16:16:59 crc kubenswrapper[4788]: E1010 16:16:59.910956 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerName="registry-server" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.910970 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerName="registry-server" Oct 10 16:16:59 crc kubenswrapper[4788]: E1010 16:16:59.911017 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerName="extract-utilities" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.911030 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerName="extract-utilities" Oct 10 16:16:59 crc kubenswrapper[4788]: E1010 16:16:59.911044 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerName="extract-content" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.911052 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerName="extract-content" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.911306 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e755581d-b00e-41d1-bb31-e4384bbf444f" containerName="mariadb-database-create" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.911326 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c86646c0-727e-4893-9d74-888fd96d2fa7" containerName="registry-server" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.912210 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9279-account-create-6hfhf" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.915565 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.921274 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9279-account-create-6hfhf"] Oct 10 16:16:59 crc kubenswrapper[4788]: I1010 16:16:59.980445 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hw7f\" (UniqueName: \"kubernetes.io/projected/578cd6ee-08cf-4a90-bb89-f9babce23451-kube-api-access-7hw7f\") pod \"glance-9279-account-create-6hfhf\" (UID: \"578cd6ee-08cf-4a90-bb89-f9babce23451\") " pod="openstack/glance-9279-account-create-6hfhf" Oct 10 16:17:00 crc kubenswrapper[4788]: I1010 16:17:00.082287 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hw7f\" (UniqueName: \"kubernetes.io/projected/578cd6ee-08cf-4a90-bb89-f9babce23451-kube-api-access-7hw7f\") pod \"glance-9279-account-create-6hfhf\" (UID: \"578cd6ee-08cf-4a90-bb89-f9babce23451\") " pod="openstack/glance-9279-account-create-6hfhf" Oct 10 16:17:00 crc kubenswrapper[4788]: I1010 16:17:00.106465 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hw7f\" (UniqueName: \"kubernetes.io/projected/578cd6ee-08cf-4a90-bb89-f9babce23451-kube-api-access-7hw7f\") pod \"glance-9279-account-create-6hfhf\" (UID: \"578cd6ee-08cf-4a90-bb89-f9babce23451\") " pod="openstack/glance-9279-account-create-6hfhf" Oct 10 16:17:00 crc kubenswrapper[4788]: I1010 16:17:00.245389 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9279-account-create-6hfhf" Oct 10 16:17:00 crc kubenswrapper[4788]: I1010 16:17:00.664871 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9279-account-create-6hfhf"] Oct 10 16:17:01 crc kubenswrapper[4788]: I1010 16:17:01.451418 4788 generic.go:334] "Generic (PLEG): container finished" podID="578cd6ee-08cf-4a90-bb89-f9babce23451" containerID="8d74f14f749a6965f4b498bdd02e32e5ff93ef766be7eccc545285f7c8808842" exitCode=0 Oct 10 16:17:01 crc kubenswrapper[4788]: I1010 16:17:01.451535 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9279-account-create-6hfhf" event={"ID":"578cd6ee-08cf-4a90-bb89-f9babce23451","Type":"ContainerDied","Data":"8d74f14f749a6965f4b498bdd02e32e5ff93ef766be7eccc545285f7c8808842"} Oct 10 16:17:01 crc kubenswrapper[4788]: I1010 16:17:01.451697 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9279-account-create-6hfhf" event={"ID":"578cd6ee-08cf-4a90-bb89-f9babce23451","Type":"ContainerStarted","Data":"aecac761528b548289942c5da547ab88f4a16aefa6053b02ecdb8cd11de2effe"} Oct 10 16:17:02 crc kubenswrapper[4788]: I1010 16:17:02.761785 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9279-account-create-6hfhf" Oct 10 16:17:02 crc kubenswrapper[4788]: I1010 16:17:02.841842 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hw7f\" (UniqueName: \"kubernetes.io/projected/578cd6ee-08cf-4a90-bb89-f9babce23451-kube-api-access-7hw7f\") pod \"578cd6ee-08cf-4a90-bb89-f9babce23451\" (UID: \"578cd6ee-08cf-4a90-bb89-f9babce23451\") " Oct 10 16:17:02 crc kubenswrapper[4788]: I1010 16:17:02.847036 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/578cd6ee-08cf-4a90-bb89-f9babce23451-kube-api-access-7hw7f" (OuterVolumeSpecName: "kube-api-access-7hw7f") pod "578cd6ee-08cf-4a90-bb89-f9babce23451" (UID: "578cd6ee-08cf-4a90-bb89-f9babce23451"). InnerVolumeSpecName "kube-api-access-7hw7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:02 crc kubenswrapper[4788]: I1010 16:17:02.944837 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hw7f\" (UniqueName: \"kubernetes.io/projected/578cd6ee-08cf-4a90-bb89-f9babce23451-kube-api-access-7hw7f\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:03 crc kubenswrapper[4788]: I1010 16:17:03.468044 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9279-account-create-6hfhf" event={"ID":"578cd6ee-08cf-4a90-bb89-f9babce23451","Type":"ContainerDied","Data":"aecac761528b548289942c5da547ab88f4a16aefa6053b02ecdb8cd11de2effe"} Oct 10 16:17:03 crc kubenswrapper[4788]: I1010 16:17:03.468303 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aecac761528b548289942c5da547ab88f4a16aefa6053b02ecdb8cd11de2effe" Oct 10 16:17:03 crc kubenswrapper[4788]: I1010 16:17:03.468122 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9279-account-create-6hfhf" Oct 10 16:17:04 crc kubenswrapper[4788]: I1010 16:17:04.959024 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-j58jf"] Oct 10 16:17:04 crc kubenswrapper[4788]: E1010 16:17:04.959436 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="578cd6ee-08cf-4a90-bb89-f9babce23451" containerName="mariadb-account-create" Oct 10 16:17:04 crc kubenswrapper[4788]: I1010 16:17:04.959452 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="578cd6ee-08cf-4a90-bb89-f9babce23451" containerName="mariadb-account-create" Oct 10 16:17:04 crc kubenswrapper[4788]: I1010 16:17:04.959679 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="578cd6ee-08cf-4a90-bb89-f9babce23451" containerName="mariadb-account-create" Oct 10 16:17:04 crc kubenswrapper[4788]: I1010 16:17:04.960332 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:04 crc kubenswrapper[4788]: I1010 16:17:04.962341 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 10 16:17:04 crc kubenswrapper[4788]: I1010 16:17:04.963802 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m6cfp" Oct 10 16:17:04 crc kubenswrapper[4788]: I1010 16:17:04.979802 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-j58jf"] Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.112341 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdvvr\" (UniqueName: \"kubernetes.io/projected/f72954d6-f07f-4272-a985-4dbcc0ae7855-kube-api-access-jdvvr\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.113333 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-db-sync-config-data\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.113621 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-combined-ca-bundle\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.114220 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-config-data\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.216555 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-config-data\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.216643 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdvvr\" (UniqueName: \"kubernetes.io/projected/f72954d6-f07f-4272-a985-4dbcc0ae7855-kube-api-access-jdvvr\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.216702 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-db-sync-config-data\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.216747 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-combined-ca-bundle\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.221682 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-config-data\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.225618 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-db-sync-config-data\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.231181 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-combined-ca-bundle\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.247264 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdvvr\" (UniqueName: \"kubernetes.io/projected/f72954d6-f07f-4272-a985-4dbcc0ae7855-kube-api-access-jdvvr\") pod \"glance-db-sync-j58jf\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.329799 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:05 crc kubenswrapper[4788]: I1010 16:17:05.876627 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-j58jf"] Oct 10 16:17:06 crc kubenswrapper[4788]: I1010 16:17:06.506901 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j58jf" event={"ID":"f72954d6-f07f-4272-a985-4dbcc0ae7855","Type":"ContainerStarted","Data":"38a1d75692076941e8b31beca56563bb4fd1cb9e8b493f305f28a2496eb5cbc2"} Oct 10 16:17:06 crc kubenswrapper[4788]: I1010 16:17:06.507413 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j58jf" event={"ID":"f72954d6-f07f-4272-a985-4dbcc0ae7855","Type":"ContainerStarted","Data":"aa21a8b0158e63a497f04de920af4fda6ac4dfcd00803d9924c058e6544a34e0"} Oct 10 16:17:06 crc kubenswrapper[4788]: I1010 16:17:06.526991 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-j58jf" podStartSLOduration=2.526973154 podStartE2EDuration="2.526973154s" podCreationTimestamp="2025-10-10 16:17:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:17:06.523686869 +0000 UTC m=+5528.973402417" watchObservedRunningTime="2025-10-10 16:17:06.526973154 +0000 UTC m=+5528.976688702" Oct 10 16:17:10 crc kubenswrapper[4788]: I1010 16:17:10.546060 4788 generic.go:334] "Generic (PLEG): container finished" podID="f72954d6-f07f-4272-a985-4dbcc0ae7855" containerID="38a1d75692076941e8b31beca56563bb4fd1cb9e8b493f305f28a2496eb5cbc2" exitCode=0 Oct 10 16:17:10 crc kubenswrapper[4788]: I1010 16:17:10.546597 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j58jf" event={"ID":"f72954d6-f07f-4272-a985-4dbcc0ae7855","Type":"ContainerDied","Data":"38a1d75692076941e8b31beca56563bb4fd1cb9e8b493f305f28a2496eb5cbc2"} Oct 10 16:17:11 crc kubenswrapper[4788]: I1010 16:17:11.941679 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.033004 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-config-data\") pod \"f72954d6-f07f-4272-a985-4dbcc0ae7855\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.033098 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-combined-ca-bundle\") pod \"f72954d6-f07f-4272-a985-4dbcc0ae7855\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.033345 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdvvr\" (UniqueName: \"kubernetes.io/projected/f72954d6-f07f-4272-a985-4dbcc0ae7855-kube-api-access-jdvvr\") pod \"f72954d6-f07f-4272-a985-4dbcc0ae7855\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.033376 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-db-sync-config-data\") pod \"f72954d6-f07f-4272-a985-4dbcc0ae7855\" (UID: \"f72954d6-f07f-4272-a985-4dbcc0ae7855\") " Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.038534 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f72954d6-f07f-4272-a985-4dbcc0ae7855" (UID: "f72954d6-f07f-4272-a985-4dbcc0ae7855"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.038694 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f72954d6-f07f-4272-a985-4dbcc0ae7855-kube-api-access-jdvvr" (OuterVolumeSpecName: "kube-api-access-jdvvr") pod "f72954d6-f07f-4272-a985-4dbcc0ae7855" (UID: "f72954d6-f07f-4272-a985-4dbcc0ae7855"). InnerVolumeSpecName "kube-api-access-jdvvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.060917 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f72954d6-f07f-4272-a985-4dbcc0ae7855" (UID: "f72954d6-f07f-4272-a985-4dbcc0ae7855"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.081346 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-config-data" (OuterVolumeSpecName: "config-data") pod "f72954d6-f07f-4272-a985-4dbcc0ae7855" (UID: "f72954d6-f07f-4272-a985-4dbcc0ae7855"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.135602 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.135632 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.135646 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdvvr\" (UniqueName: \"kubernetes.io/projected/f72954d6-f07f-4272-a985-4dbcc0ae7855-kube-api-access-jdvvr\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.135657 4788 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f72954d6-f07f-4272-a985-4dbcc0ae7855-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.572477 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j58jf" event={"ID":"f72954d6-f07f-4272-a985-4dbcc0ae7855","Type":"ContainerDied","Data":"aa21a8b0158e63a497f04de920af4fda6ac4dfcd00803d9924c058e6544a34e0"} Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.572537 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa21a8b0158e63a497f04de920af4fda6ac4dfcd00803d9924c058e6544a34e0" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.572557 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j58jf" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.673112 4788 scope.go:117] "RemoveContainer" containerID="1113508caceb0d522af8209c17bf099d5e088853b3299f5a3c49843e65d5d1e9" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.917721 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:12 crc kubenswrapper[4788]: E1010 16:17:12.918496 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f72954d6-f07f-4272-a985-4dbcc0ae7855" containerName="glance-db-sync" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.918516 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f72954d6-f07f-4272-a985-4dbcc0ae7855" containerName="glance-db-sync" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.918704 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f72954d6-f07f-4272-a985-4dbcc0ae7855" containerName="glance-db-sync" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.920452 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.925329 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-m6cfp" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.925616 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.925745 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.926019 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 16:17:12 crc kubenswrapper[4788]: I1010 16:17:12.935804 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.007745 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f8774dc7-5jnft"] Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.009681 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.018272 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f8774dc7-5jnft"] Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.068436 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.068509 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.068548 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.068632 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5xv2\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-kube-api-access-z5xv2\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.068668 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.068717 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.068855 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-logs\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.170840 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.170910 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.170953 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-sb\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.170992 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7gbm\" (UniqueName: \"kubernetes.io/projected/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-kube-api-access-r7gbm\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171015 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5xv2\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-kube-api-access-z5xv2\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171036 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171215 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171315 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-logs\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171352 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-config\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171426 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-dns-svc\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171479 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-nb\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171511 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.171508 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.172027 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-logs\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.176691 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.176860 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.177464 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.183963 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.196627 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.196748 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5xv2\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-kube-api-access-z5xv2\") pod \"glance-default-external-api-0\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.198129 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.203652 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.219637 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.248334 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.274351 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-config\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.274409 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-dns-svc\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.274442 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-nb\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.274521 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-sb\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.274556 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7gbm\" (UniqueName: \"kubernetes.io/projected/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-kube-api-access-r7gbm\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.275715 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-config\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.276027 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-dns-svc\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.276046 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-nb\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.276706 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-sb\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.306413 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7gbm\" (UniqueName: \"kubernetes.io/projected/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-kube-api-access-r7gbm\") pod \"dnsmasq-dns-5f8774dc7-5jnft\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.335892 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.376319 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.376359 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-logs\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.376397 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.376505 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.376523 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.376542 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn6zd\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-kube-api-access-gn6zd\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.376564 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.478710 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.478759 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-logs\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.478785 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.478875 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.478896 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.478919 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn6zd\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-kube-api-access-gn6zd\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.478968 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.480626 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.480665 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-logs\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.483904 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.484884 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.491734 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.492269 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.495969 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn6zd\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-kube-api-access-gn6zd\") pod \"glance-default-internal-api-0\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.657494 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.719129 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:13 crc kubenswrapper[4788]: I1010 16:17:13.919816 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f8774dc7-5jnft"] Oct 10 16:17:14 crc kubenswrapper[4788]: I1010 16:17:14.344491 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:14 crc kubenswrapper[4788]: W1010 16:17:14.377853 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6403a351_eb70_4985_8e95_4f77bf11f64b.slice/crio-e257bda072b2d5fee96a291e18b3a82baa2de792f26e7c43ebc936b0e7aec1fc WatchSource:0}: Error finding container e257bda072b2d5fee96a291e18b3a82baa2de792f26e7c43ebc936b0e7aec1fc: Status 404 returned error can't find the container with id e257bda072b2d5fee96a291e18b3a82baa2de792f26e7c43ebc936b0e7aec1fc Oct 10 16:17:14 crc kubenswrapper[4788]: I1010 16:17:14.544411 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:14 crc kubenswrapper[4788]: I1010 16:17:14.597429 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6403a351-eb70-4985-8e95-4f77bf11f64b","Type":"ContainerStarted","Data":"e257bda072b2d5fee96a291e18b3a82baa2de792f26e7c43ebc936b0e7aec1fc"} Oct 10 16:17:14 crc kubenswrapper[4788]: I1010 16:17:14.601640 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a462c883-6443-498f-9bad-1ef15b0976eb","Type":"ContainerStarted","Data":"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d"} Oct 10 16:17:14 crc kubenswrapper[4788]: I1010 16:17:14.601705 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a462c883-6443-498f-9bad-1ef15b0976eb","Type":"ContainerStarted","Data":"39cf1513010fc0be23864fe925ade439c9f58fd25f41f01fa7422651b2745472"} Oct 10 16:17:14 crc kubenswrapper[4788]: I1010 16:17:14.606406 4788 generic.go:334] "Generic (PLEG): container finished" podID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" containerID="bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b" exitCode=0 Oct 10 16:17:14 crc kubenswrapper[4788]: I1010 16:17:14.606462 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" event={"ID":"8e9df50e-c08f-4bf7-ad94-22813cc9efa6","Type":"ContainerDied","Data":"bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b"} Oct 10 16:17:14 crc kubenswrapper[4788]: I1010 16:17:14.606513 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" event={"ID":"8e9df50e-c08f-4bf7-ad94-22813cc9efa6","Type":"ContainerStarted","Data":"f2e969910fc76eb2320e1348ca4178d8f5ebeb3ab4055f0d63a6a8c3ec059caf"} Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.616287 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" event={"ID":"8e9df50e-c08f-4bf7-ad94-22813cc9efa6","Type":"ContainerStarted","Data":"85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41"} Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.618417 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.618616 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6403a351-eb70-4985-8e95-4f77bf11f64b","Type":"ContainerStarted","Data":"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1"} Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.618658 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6403a351-eb70-4985-8e95-4f77bf11f64b","Type":"ContainerStarted","Data":"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8"} Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.621355 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a462c883-6443-498f-9bad-1ef15b0976eb","Type":"ContainerStarted","Data":"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253"} Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.621431 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" containerName="glance-log" containerID="cri-o://79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d" gracePeriod=30 Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.621541 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" containerName="glance-httpd" containerID="cri-o://ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253" gracePeriod=30 Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.644742 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" podStartSLOduration=3.644722396 podStartE2EDuration="3.644722396s" podCreationTimestamp="2025-10-10 16:17:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:17:15.641546055 +0000 UTC m=+5538.091261613" watchObservedRunningTime="2025-10-10 16:17:15.644722396 +0000 UTC m=+5538.094437944" Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.667842 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.66781884 podStartE2EDuration="3.66781884s" podCreationTimestamp="2025-10-10 16:17:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:17:15.658031889 +0000 UTC m=+5538.107747447" watchObservedRunningTime="2025-10-10 16:17:15.66781884 +0000 UTC m=+5538.117534398" Oct 10 16:17:15 crc kubenswrapper[4788]: I1010 16:17:15.716032 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.716010898 podStartE2EDuration="2.716010898s" podCreationTimestamp="2025-10-10 16:17:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:17:15.701943786 +0000 UTC m=+5538.151659334" watchObservedRunningTime="2025-10-10 16:17:15.716010898 +0000 UTC m=+5538.165726446" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.293109 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.343481 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5xv2\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-kube-api-access-z5xv2\") pod \"a462c883-6443-498f-9bad-1ef15b0976eb\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.343827 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-logs\") pod \"a462c883-6443-498f-9bad-1ef15b0976eb\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.343916 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-scripts\") pod \"a462c883-6443-498f-9bad-1ef15b0976eb\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.343971 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-httpd-run\") pod \"a462c883-6443-498f-9bad-1ef15b0976eb\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.343993 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-combined-ca-bundle\") pod \"a462c883-6443-498f-9bad-1ef15b0976eb\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.344032 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-config-data\") pod \"a462c883-6443-498f-9bad-1ef15b0976eb\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.344123 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-ceph\") pod \"a462c883-6443-498f-9bad-1ef15b0976eb\" (UID: \"a462c883-6443-498f-9bad-1ef15b0976eb\") " Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.350826 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a462c883-6443-498f-9bad-1ef15b0976eb" (UID: "a462c883-6443-498f-9bad-1ef15b0976eb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.351010 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-ceph" (OuterVolumeSpecName: "ceph") pod "a462c883-6443-498f-9bad-1ef15b0976eb" (UID: "a462c883-6443-498f-9bad-1ef15b0976eb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.351530 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-kube-api-access-z5xv2" (OuterVolumeSpecName: "kube-api-access-z5xv2") pod "a462c883-6443-498f-9bad-1ef15b0976eb" (UID: "a462c883-6443-498f-9bad-1ef15b0976eb"). InnerVolumeSpecName "kube-api-access-z5xv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.356244 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-scripts" (OuterVolumeSpecName: "scripts") pod "a462c883-6443-498f-9bad-1ef15b0976eb" (UID: "a462c883-6443-498f-9bad-1ef15b0976eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.357069 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-logs" (OuterVolumeSpecName: "logs") pod "a462c883-6443-498f-9bad-1ef15b0976eb" (UID: "a462c883-6443-498f-9bad-1ef15b0976eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.383462 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a462c883-6443-498f-9bad-1ef15b0976eb" (UID: "a462c883-6443-498f-9bad-1ef15b0976eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.419359 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-config-data" (OuterVolumeSpecName: "config-data") pod "a462c883-6443-498f-9bad-1ef15b0976eb" (UID: "a462c883-6443-498f-9bad-1ef15b0976eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.446677 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.446709 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.446719 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a462c883-6443-498f-9bad-1ef15b0976eb-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.446728 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.446740 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a462c883-6443-498f-9bad-1ef15b0976eb-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.446747 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.446756 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5xv2\" (UniqueName: \"kubernetes.io/projected/a462c883-6443-498f-9bad-1ef15b0976eb-kube-api-access-z5xv2\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.632605 4788 generic.go:334] "Generic (PLEG): container finished" podID="a462c883-6443-498f-9bad-1ef15b0976eb" containerID="ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253" exitCode=0 Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.632646 4788 generic.go:334] "Generic (PLEG): container finished" podID="a462c883-6443-498f-9bad-1ef15b0976eb" containerID="79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d" exitCode=143 Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.632705 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.632716 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a462c883-6443-498f-9bad-1ef15b0976eb","Type":"ContainerDied","Data":"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253"} Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.632776 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a462c883-6443-498f-9bad-1ef15b0976eb","Type":"ContainerDied","Data":"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d"} Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.632788 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a462c883-6443-498f-9bad-1ef15b0976eb","Type":"ContainerDied","Data":"39cf1513010fc0be23864fe925ade439c9f58fd25f41f01fa7422651b2745472"} Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.632804 4788 scope.go:117] "RemoveContainer" containerID="ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.658966 4788 scope.go:117] "RemoveContainer" containerID="79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.672122 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.681751 4788 scope.go:117] "RemoveContainer" containerID="ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253" Oct 10 16:17:16 crc kubenswrapper[4788]: E1010 16:17:16.682932 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253\": container with ID starting with ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253 not found: ID does not exist" containerID="ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.682968 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253"} err="failed to get container status \"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253\": rpc error: code = NotFound desc = could not find container \"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253\": container with ID starting with ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253 not found: ID does not exist" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.682993 4788 scope.go:117] "RemoveContainer" containerID="79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d" Oct 10 16:17:16 crc kubenswrapper[4788]: E1010 16:17:16.684320 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d\": container with ID starting with 79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d not found: ID does not exist" containerID="79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.684347 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d"} err="failed to get container status \"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d\": rpc error: code = NotFound desc = could not find container \"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d\": container with ID starting with 79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d not found: ID does not exist" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.684363 4788 scope.go:117] "RemoveContainer" containerID="ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.684611 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253"} err="failed to get container status \"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253\": rpc error: code = NotFound desc = could not find container \"ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253\": container with ID starting with ac063f4c3889260e48818e6acf3ca15992923ca88fe08e180b42137b0fb49253 not found: ID does not exist" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.684651 4788 scope.go:117] "RemoveContainer" containerID="79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.684864 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d"} err="failed to get container status \"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d\": rpc error: code = NotFound desc = could not find container \"79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d\": container with ID starting with 79b4a5453299359fbd0f0c340bfe3145a0362b0dcb0e1735fdb6fc5574f7cc4d not found: ID does not exist" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.685696 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.704812 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:16 crc kubenswrapper[4788]: E1010 16:17:16.705181 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" containerName="glance-log" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.705193 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" containerName="glance-log" Oct 10 16:17:16 crc kubenswrapper[4788]: E1010 16:17:16.705231 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" containerName="glance-httpd" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.705237 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" containerName="glance-httpd" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.705395 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" containerName="glance-log" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.705406 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" containerName="glance-httpd" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.706295 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.709333 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.720676 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.852772 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.852933 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.852999 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.853053 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-ceph\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.853076 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq7s9\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-kube-api-access-kq7s9\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.853123 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.853269 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-logs\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.955254 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.955366 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.955437 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.955468 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-ceph\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.955655 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq7s9\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-kube-api-access-kq7s9\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.955955 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.956000 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-logs\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.956019 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.956345 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-logs\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.960977 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-ceph\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.960994 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.965252 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.966619 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:16 crc kubenswrapper[4788]: I1010 16:17:16.991545 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq7s9\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-kube-api-access-kq7s9\") pod \"glance-default-external-api-0\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " pod="openstack/glance-default-external-api-0" Oct 10 16:17:17 crc kubenswrapper[4788]: I1010 16:17:17.025579 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:17:17 crc kubenswrapper[4788]: I1010 16:17:17.124153 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:17 crc kubenswrapper[4788]: I1010 16:17:17.607716 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:17:17 crc kubenswrapper[4788]: W1010 16:17:17.613424 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12d32477_f9af_47c5_9b91_0ab82d677e5a.slice/crio-93db2822c04ee7d25781f3dbe070716cdd7ab2920a28fd60f505b14c397f3d21 WatchSource:0}: Error finding container 93db2822c04ee7d25781f3dbe070716cdd7ab2920a28fd60f505b14c397f3d21: Status 404 returned error can't find the container with id 93db2822c04ee7d25781f3dbe070716cdd7ab2920a28fd60f505b14c397f3d21 Oct 10 16:17:17 crc kubenswrapper[4788]: I1010 16:17:17.643832 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12d32477-f9af-47c5-9b91-0ab82d677e5a","Type":"ContainerStarted","Data":"93db2822c04ee7d25781f3dbe070716cdd7ab2920a28fd60f505b14c397f3d21"} Oct 10 16:17:17 crc kubenswrapper[4788]: I1010 16:17:17.645574 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerName="glance-log" containerID="cri-o://442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8" gracePeriod=30 Oct 10 16:17:17 crc kubenswrapper[4788]: I1010 16:17:17.645654 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerName="glance-httpd" containerID="cri-o://1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1" gracePeriod=30 Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.250695 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a462c883-6443-498f-9bad-1ef15b0976eb" path="/var/lib/kubelet/pods/a462c883-6443-498f-9bad-1ef15b0976eb/volumes" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.273842 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.379880 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-scripts\") pod \"6403a351-eb70-4985-8e95-4f77bf11f64b\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.380038 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-config-data\") pod \"6403a351-eb70-4985-8e95-4f77bf11f64b\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.380068 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-combined-ca-bundle\") pod \"6403a351-eb70-4985-8e95-4f77bf11f64b\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.380102 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-httpd-run\") pod \"6403a351-eb70-4985-8e95-4f77bf11f64b\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.380248 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-logs\") pod \"6403a351-eb70-4985-8e95-4f77bf11f64b\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.380271 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-ceph\") pod \"6403a351-eb70-4985-8e95-4f77bf11f64b\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.380289 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn6zd\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-kube-api-access-gn6zd\") pod \"6403a351-eb70-4985-8e95-4f77bf11f64b\" (UID: \"6403a351-eb70-4985-8e95-4f77bf11f64b\") " Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.380753 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-logs" (OuterVolumeSpecName: "logs") pod "6403a351-eb70-4985-8e95-4f77bf11f64b" (UID: "6403a351-eb70-4985-8e95-4f77bf11f64b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.380839 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6403a351-eb70-4985-8e95-4f77bf11f64b" (UID: "6403a351-eb70-4985-8e95-4f77bf11f64b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.381030 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.381050 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6403a351-eb70-4985-8e95-4f77bf11f64b-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.384565 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-kube-api-access-gn6zd" (OuterVolumeSpecName: "kube-api-access-gn6zd") pod "6403a351-eb70-4985-8e95-4f77bf11f64b" (UID: "6403a351-eb70-4985-8e95-4f77bf11f64b"). InnerVolumeSpecName "kube-api-access-gn6zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.384891 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-ceph" (OuterVolumeSpecName: "ceph") pod "6403a351-eb70-4985-8e95-4f77bf11f64b" (UID: "6403a351-eb70-4985-8e95-4f77bf11f64b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.392807 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-scripts" (OuterVolumeSpecName: "scripts") pod "6403a351-eb70-4985-8e95-4f77bf11f64b" (UID: "6403a351-eb70-4985-8e95-4f77bf11f64b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.406688 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6403a351-eb70-4985-8e95-4f77bf11f64b" (UID: "6403a351-eb70-4985-8e95-4f77bf11f64b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.432868 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-config-data" (OuterVolumeSpecName: "config-data") pod "6403a351-eb70-4985-8e95-4f77bf11f64b" (UID: "6403a351-eb70-4985-8e95-4f77bf11f64b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.483424 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.483627 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.483647 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6403a351-eb70-4985-8e95-4f77bf11f64b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.483663 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.483674 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn6zd\" (UniqueName: \"kubernetes.io/projected/6403a351-eb70-4985-8e95-4f77bf11f64b-kube-api-access-gn6zd\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.661866 4788 generic.go:334] "Generic (PLEG): container finished" podID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerID="1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1" exitCode=0 Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.661919 4788 generic.go:334] "Generic (PLEG): container finished" podID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerID="442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8" exitCode=143 Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.662058 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6403a351-eb70-4985-8e95-4f77bf11f64b","Type":"ContainerDied","Data":"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1"} Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.662095 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6403a351-eb70-4985-8e95-4f77bf11f64b","Type":"ContainerDied","Data":"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8"} Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.662152 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.662107 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6403a351-eb70-4985-8e95-4f77bf11f64b","Type":"ContainerDied","Data":"e257bda072b2d5fee96a291e18b3a82baa2de792f26e7c43ebc936b0e7aec1fc"} Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.662223 4788 scope.go:117] "RemoveContainer" containerID="1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.664899 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12d32477-f9af-47c5-9b91-0ab82d677e5a","Type":"ContainerStarted","Data":"ea00096efb6dc90a7bad6d9e9a7da803eccf8477e55a62dc236a87f44aa69e1e"} Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.685168 4788 scope.go:117] "RemoveContainer" containerID="442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.697325 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.697305086 podStartE2EDuration="2.697305086s" podCreationTimestamp="2025-10-10 16:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:17:18.691298152 +0000 UTC m=+5541.141013700" watchObservedRunningTime="2025-10-10 16:17:18.697305086 +0000 UTC m=+5541.147020634" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.721414 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.729399 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.732976 4788 scope.go:117] "RemoveContainer" containerID="1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1" Oct 10 16:17:18 crc kubenswrapper[4788]: E1010 16:17:18.738363 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1\": container with ID starting with 1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1 not found: ID does not exist" containerID="1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.738433 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1"} err="failed to get container status \"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1\": rpc error: code = NotFound desc = could not find container \"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1\": container with ID starting with 1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1 not found: ID does not exist" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.738460 4788 scope.go:117] "RemoveContainer" containerID="442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8" Oct 10 16:17:18 crc kubenswrapper[4788]: E1010 16:17:18.741508 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8\": container with ID starting with 442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8 not found: ID does not exist" containerID="442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.741556 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8"} err="failed to get container status \"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8\": rpc error: code = NotFound desc = could not find container \"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8\": container with ID starting with 442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8 not found: ID does not exist" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.741592 4788 scope.go:117] "RemoveContainer" containerID="1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.742093 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1"} err="failed to get container status \"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1\": rpc error: code = NotFound desc = could not find container \"1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1\": container with ID starting with 1154c954c3bfbb996d5d6e8b4b15f0268946e42f15f9e5e70abf8e6164451db1 not found: ID does not exist" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.742149 4788 scope.go:117] "RemoveContainer" containerID="442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.742553 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8"} err="failed to get container status \"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8\": rpc error: code = NotFound desc = could not find container \"442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8\": container with ID starting with 442080a764e04b863bb2d32630c01d179361332e7559f7a56fc8b79d2f45c5b8 not found: ID does not exist" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.751563 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:18 crc kubenswrapper[4788]: E1010 16:17:18.751979 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerName="glance-log" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.751992 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerName="glance-log" Oct 10 16:17:18 crc kubenswrapper[4788]: E1010 16:17:18.752020 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerName="glance-httpd" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.752026 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerName="glance-httpd" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.752220 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerName="glance-httpd" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.752241 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" containerName="glance-log" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.754233 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.761177 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.766190 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.895461 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g2hf\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-kube-api-access-9g2hf\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.895856 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.895975 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.896002 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.896071 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.896103 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.896131 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-logs\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.998011 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g2hf\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-kube-api-access-9g2hf\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.998103 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.998124 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.998169 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.998238 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.998258 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.998277 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-logs\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.999019 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:18 crc kubenswrapper[4788]: I1010 16:17:18.999049 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-logs\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:19 crc kubenswrapper[4788]: I1010 16:17:19.003718 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:19 crc kubenswrapper[4788]: I1010 16:17:19.003776 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:19 crc kubenswrapper[4788]: I1010 16:17:19.005406 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:19 crc kubenswrapper[4788]: I1010 16:17:19.008060 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:19 crc kubenswrapper[4788]: I1010 16:17:19.014240 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g2hf\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-kube-api-access-9g2hf\") pod \"glance-default-internal-api-0\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:17:19 crc kubenswrapper[4788]: I1010 16:17:19.095378 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:19 crc kubenswrapper[4788]: I1010 16:17:19.659327 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:17:19 crc kubenswrapper[4788]: W1010 16:17:19.665006 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce16116b_c7d1_4d82_ab72_34002a0dbecc.slice/crio-399b0465c36ca0ae9ee6c85416552c8749fc16ef3d88b765317e7f62a8acabe6 WatchSource:0}: Error finding container 399b0465c36ca0ae9ee6c85416552c8749fc16ef3d88b765317e7f62a8acabe6: Status 404 returned error can't find the container with id 399b0465c36ca0ae9ee6c85416552c8749fc16ef3d88b765317e7f62a8acabe6 Oct 10 16:17:19 crc kubenswrapper[4788]: I1010 16:17:19.681786 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12d32477-f9af-47c5-9b91-0ab82d677e5a","Type":"ContainerStarted","Data":"8c13c7adb5a1a8363332b520e23b991f39b08314bc744aa9f43ca67d699afda5"} Oct 10 16:17:20 crc kubenswrapper[4788]: I1010 16:17:20.243863 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6403a351-eb70-4985-8e95-4f77bf11f64b" path="/var/lib/kubelet/pods/6403a351-eb70-4985-8e95-4f77bf11f64b/volumes" Oct 10 16:17:20 crc kubenswrapper[4788]: I1010 16:17:20.693222 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ce16116b-c7d1-4d82-ab72-34002a0dbecc","Type":"ContainerStarted","Data":"7eae1f41b35d6cd12e721bea2a731f1610c78b398434a5a866d429fe7716e144"} Oct 10 16:17:20 crc kubenswrapper[4788]: I1010 16:17:20.693268 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ce16116b-c7d1-4d82-ab72-34002a0dbecc","Type":"ContainerStarted","Data":"39c96d7e490e032971909cbaf3a95df7e21aaf7fa95cc28e68d4cb76fc8a5c7a"} Oct 10 16:17:20 crc kubenswrapper[4788]: I1010 16:17:20.693278 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ce16116b-c7d1-4d82-ab72-34002a0dbecc","Type":"ContainerStarted","Data":"399b0465c36ca0ae9ee6c85416552c8749fc16ef3d88b765317e7f62a8acabe6"} Oct 10 16:17:20 crc kubenswrapper[4788]: I1010 16:17:20.712921 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.712902357 podStartE2EDuration="2.712902357s" podCreationTimestamp="2025-10-10 16:17:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:17:20.708636667 +0000 UTC m=+5543.158352215" watchObservedRunningTime="2025-10-10 16:17:20.712902357 +0000 UTC m=+5543.162617925" Oct 10 16:17:23 crc kubenswrapper[4788]: I1010 16:17:23.365279 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:17:23 crc kubenswrapper[4788]: I1010 16:17:23.435917 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d6f69859c-knp8p"] Oct 10 16:17:23 crc kubenswrapper[4788]: I1010 16:17:23.436280 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" podUID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" containerName="dnsmasq-dns" containerID="cri-o://3232984f0d4d5643f20432f86faa2c36415b2693817ef1b7fa65447d79b458d6" gracePeriod=10 Oct 10 16:17:23 crc kubenswrapper[4788]: I1010 16:17:23.732623 4788 generic.go:334] "Generic (PLEG): container finished" podID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" containerID="3232984f0d4d5643f20432f86faa2c36415b2693817ef1b7fa65447d79b458d6" exitCode=0 Oct 10 16:17:23 crc kubenswrapper[4788]: I1010 16:17:23.732668 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" event={"ID":"68cde0e7-40e4-4f5a-a6f3-50d10996a32e","Type":"ContainerDied","Data":"3232984f0d4d5643f20432f86faa2c36415b2693817ef1b7fa65447d79b458d6"} Oct 10 16:17:23 crc kubenswrapper[4788]: I1010 16:17:23.945639 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.029215 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-config\") pod \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.029303 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-nb\") pod \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.029349 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb965\" (UniqueName: \"kubernetes.io/projected/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-kube-api-access-tb965\") pod \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.029375 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-sb\") pod \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.029460 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-dns-svc\") pod \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\" (UID: \"68cde0e7-40e4-4f5a-a6f3-50d10996a32e\") " Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.036382 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-kube-api-access-tb965" (OuterVolumeSpecName: "kube-api-access-tb965") pod "68cde0e7-40e4-4f5a-a6f3-50d10996a32e" (UID: "68cde0e7-40e4-4f5a-a6f3-50d10996a32e"). InnerVolumeSpecName "kube-api-access-tb965". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.068984 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "68cde0e7-40e4-4f5a-a6f3-50d10996a32e" (UID: "68cde0e7-40e4-4f5a-a6f3-50d10996a32e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.077911 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68cde0e7-40e4-4f5a-a6f3-50d10996a32e" (UID: "68cde0e7-40e4-4f5a-a6f3-50d10996a32e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.079025 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-config" (OuterVolumeSpecName: "config") pod "68cde0e7-40e4-4f5a-a6f3-50d10996a32e" (UID: "68cde0e7-40e4-4f5a-a6f3-50d10996a32e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.081130 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "68cde0e7-40e4-4f5a-a6f3-50d10996a32e" (UID: "68cde0e7-40e4-4f5a-a6f3-50d10996a32e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.131536 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.131584 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb965\" (UniqueName: \"kubernetes.io/projected/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-kube-api-access-tb965\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.131596 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.131605 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.131613 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68cde0e7-40e4-4f5a-a6f3-50d10996a32e-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.744363 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" event={"ID":"68cde0e7-40e4-4f5a-a6f3-50d10996a32e","Type":"ContainerDied","Data":"afde627d7650c7d23a05918a7006b5914815db5fe44ce4890f9284a5456242d0"} Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.744413 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6f69859c-knp8p" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.744651 4788 scope.go:117] "RemoveContainer" containerID="3232984f0d4d5643f20432f86faa2c36415b2693817ef1b7fa65447d79b458d6" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.765258 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d6f69859c-knp8p"] Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.765663 4788 scope.go:117] "RemoveContainer" containerID="73cb1240d9c3acf5bde0bafff031bafedcf5abe442c9ae2f03c6698583940898" Oct 10 16:17:24 crc kubenswrapper[4788]: I1010 16:17:24.776847 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d6f69859c-knp8p"] Oct 10 16:17:26 crc kubenswrapper[4788]: I1010 16:17:26.248737 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" path="/var/lib/kubelet/pods/68cde0e7-40e4-4f5a-a6f3-50d10996a32e/volumes" Oct 10 16:17:27 crc kubenswrapper[4788]: I1010 16:17:27.027262 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 16:17:27 crc kubenswrapper[4788]: I1010 16:17:27.027315 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 16:17:27 crc kubenswrapper[4788]: I1010 16:17:27.062889 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 16:17:27 crc kubenswrapper[4788]: I1010 16:17:27.094946 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 16:17:27 crc kubenswrapper[4788]: I1010 16:17:27.771385 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 16:17:27 crc kubenswrapper[4788]: I1010 16:17:27.772233 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.095939 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.096243 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.130296 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.147293 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.406341 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.406418 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.792779 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.792824 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.801253 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.801348 4788 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 16:17:29 crc kubenswrapper[4788]: I1010 16:17:29.803280 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 16:17:31 crc kubenswrapper[4788]: I1010 16:17:31.993091 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:31 crc kubenswrapper[4788]: I1010 16:17:31.994059 4788 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 16:17:31 crc kubenswrapper[4788]: I1010 16:17:31.999273 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.557097 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-2brsx"] Oct 10 16:17:40 crc kubenswrapper[4788]: E1010 16:17:40.558348 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" containerName="init" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.558363 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" containerName="init" Oct 10 16:17:40 crc kubenswrapper[4788]: E1010 16:17:40.558384 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" containerName="dnsmasq-dns" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.558395 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" containerName="dnsmasq-dns" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.558636 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="68cde0e7-40e4-4f5a-a6f3-50d10996a32e" containerName="dnsmasq-dns" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.559544 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2brsx" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.572233 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2brsx"] Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.643244 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbg4l\" (UniqueName: \"kubernetes.io/projected/0e49ffce-869d-4edc-ac30-80ce5040803c-kube-api-access-dbg4l\") pod \"placement-db-create-2brsx\" (UID: \"0e49ffce-869d-4edc-ac30-80ce5040803c\") " pod="openstack/placement-db-create-2brsx" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.745340 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbg4l\" (UniqueName: \"kubernetes.io/projected/0e49ffce-869d-4edc-ac30-80ce5040803c-kube-api-access-dbg4l\") pod \"placement-db-create-2brsx\" (UID: \"0e49ffce-869d-4edc-ac30-80ce5040803c\") " pod="openstack/placement-db-create-2brsx" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.768260 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbg4l\" (UniqueName: \"kubernetes.io/projected/0e49ffce-869d-4edc-ac30-80ce5040803c-kube-api-access-dbg4l\") pod \"placement-db-create-2brsx\" (UID: \"0e49ffce-869d-4edc-ac30-80ce5040803c\") " pod="openstack/placement-db-create-2brsx" Oct 10 16:17:40 crc kubenswrapper[4788]: I1010 16:17:40.884893 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2brsx" Oct 10 16:17:41 crc kubenswrapper[4788]: I1010 16:17:41.348439 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2brsx"] Oct 10 16:17:41 crc kubenswrapper[4788]: W1010 16:17:41.354087 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e49ffce_869d_4edc_ac30_80ce5040803c.slice/crio-e6cbcd683f23f4166d88c7cb5bb50474c9fb04b08908a414b486549b9ff0b4d1 WatchSource:0}: Error finding container e6cbcd683f23f4166d88c7cb5bb50474c9fb04b08908a414b486549b9ff0b4d1: Status 404 returned error can't find the container with id e6cbcd683f23f4166d88c7cb5bb50474c9fb04b08908a414b486549b9ff0b4d1 Oct 10 16:17:41 crc kubenswrapper[4788]: I1010 16:17:41.921525 4788 generic.go:334] "Generic (PLEG): container finished" podID="0e49ffce-869d-4edc-ac30-80ce5040803c" containerID="f7badbc58fd5f032a6b5505500f9bcdf2769a462dfe938366f55383bd7f1877f" exitCode=0 Oct 10 16:17:41 crc kubenswrapper[4788]: I1010 16:17:41.921611 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2brsx" event={"ID":"0e49ffce-869d-4edc-ac30-80ce5040803c","Type":"ContainerDied","Data":"f7badbc58fd5f032a6b5505500f9bcdf2769a462dfe938366f55383bd7f1877f"} Oct 10 16:17:41 crc kubenswrapper[4788]: I1010 16:17:41.922055 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2brsx" event={"ID":"0e49ffce-869d-4edc-ac30-80ce5040803c","Type":"ContainerStarted","Data":"e6cbcd683f23f4166d88c7cb5bb50474c9fb04b08908a414b486549b9ff0b4d1"} Oct 10 16:17:43 crc kubenswrapper[4788]: I1010 16:17:43.388122 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2brsx" Oct 10 16:17:43 crc kubenswrapper[4788]: I1010 16:17:43.509103 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbg4l\" (UniqueName: \"kubernetes.io/projected/0e49ffce-869d-4edc-ac30-80ce5040803c-kube-api-access-dbg4l\") pod \"0e49ffce-869d-4edc-ac30-80ce5040803c\" (UID: \"0e49ffce-869d-4edc-ac30-80ce5040803c\") " Oct 10 16:17:43 crc kubenswrapper[4788]: I1010 16:17:43.515410 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e49ffce-869d-4edc-ac30-80ce5040803c-kube-api-access-dbg4l" (OuterVolumeSpecName: "kube-api-access-dbg4l") pod "0e49ffce-869d-4edc-ac30-80ce5040803c" (UID: "0e49ffce-869d-4edc-ac30-80ce5040803c"). InnerVolumeSpecName "kube-api-access-dbg4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:43 crc kubenswrapper[4788]: I1010 16:17:43.611643 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbg4l\" (UniqueName: \"kubernetes.io/projected/0e49ffce-869d-4edc-ac30-80ce5040803c-kube-api-access-dbg4l\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:43 crc kubenswrapper[4788]: I1010 16:17:43.947814 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2brsx" event={"ID":"0e49ffce-869d-4edc-ac30-80ce5040803c","Type":"ContainerDied","Data":"e6cbcd683f23f4166d88c7cb5bb50474c9fb04b08908a414b486549b9ff0b4d1"} Oct 10 16:17:43 crc kubenswrapper[4788]: I1010 16:17:43.948085 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6cbcd683f23f4166d88c7cb5bb50474c9fb04b08908a414b486549b9ff0b4d1" Oct 10 16:17:43 crc kubenswrapper[4788]: I1010 16:17:43.947894 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2brsx" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.613611 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c7cb-account-create-v5t9s"] Oct 10 16:17:50 crc kubenswrapper[4788]: E1010 16:17:50.614797 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e49ffce-869d-4edc-ac30-80ce5040803c" containerName="mariadb-database-create" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.614815 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e49ffce-869d-4edc-ac30-80ce5040803c" containerName="mariadb-database-create" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.615035 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e49ffce-869d-4edc-ac30-80ce5040803c" containerName="mariadb-database-create" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.615866 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c7cb-account-create-v5t9s" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.621063 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.622936 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c7cb-account-create-v5t9s"] Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.667061 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrzj2\" (UniqueName: \"kubernetes.io/projected/f2cd6400-29f0-41da-9c22-0443d88dca9d-kube-api-access-qrzj2\") pod \"placement-c7cb-account-create-v5t9s\" (UID: \"f2cd6400-29f0-41da-9c22-0443d88dca9d\") " pod="openstack/placement-c7cb-account-create-v5t9s" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.769334 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrzj2\" (UniqueName: \"kubernetes.io/projected/f2cd6400-29f0-41da-9c22-0443d88dca9d-kube-api-access-qrzj2\") pod \"placement-c7cb-account-create-v5t9s\" (UID: \"f2cd6400-29f0-41da-9c22-0443d88dca9d\") " pod="openstack/placement-c7cb-account-create-v5t9s" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.799618 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrzj2\" (UniqueName: \"kubernetes.io/projected/f2cd6400-29f0-41da-9c22-0443d88dca9d-kube-api-access-qrzj2\") pod \"placement-c7cb-account-create-v5t9s\" (UID: \"f2cd6400-29f0-41da-9c22-0443d88dca9d\") " pod="openstack/placement-c7cb-account-create-v5t9s" Oct 10 16:17:50 crc kubenswrapper[4788]: I1010 16:17:50.944226 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c7cb-account-create-v5t9s" Oct 10 16:17:51 crc kubenswrapper[4788]: I1010 16:17:51.484391 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c7cb-account-create-v5t9s"] Oct 10 16:17:52 crc kubenswrapper[4788]: I1010 16:17:52.028889 4788 generic.go:334] "Generic (PLEG): container finished" podID="f2cd6400-29f0-41da-9c22-0443d88dca9d" containerID="c891a5dfbf246674ff54c6a3fb9e32fd9ad162df65a95aeb0f639bfed9af6f87" exitCode=0 Oct 10 16:17:52 crc kubenswrapper[4788]: I1010 16:17:52.029122 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c7cb-account-create-v5t9s" event={"ID":"f2cd6400-29f0-41da-9c22-0443d88dca9d","Type":"ContainerDied","Data":"c891a5dfbf246674ff54c6a3fb9e32fd9ad162df65a95aeb0f639bfed9af6f87"} Oct 10 16:17:52 crc kubenswrapper[4788]: I1010 16:17:52.029250 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c7cb-account-create-v5t9s" event={"ID":"f2cd6400-29f0-41da-9c22-0443d88dca9d","Type":"ContainerStarted","Data":"be7fe79464afa39c1c85a2261047c9cc8df9a471a9381cdcfb11fa7f8d57048f"} Oct 10 16:17:53 crc kubenswrapper[4788]: I1010 16:17:53.371440 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c7cb-account-create-v5t9s" Oct 10 16:17:53 crc kubenswrapper[4788]: I1010 16:17:53.416362 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrzj2\" (UniqueName: \"kubernetes.io/projected/f2cd6400-29f0-41da-9c22-0443d88dca9d-kube-api-access-qrzj2\") pod \"f2cd6400-29f0-41da-9c22-0443d88dca9d\" (UID: \"f2cd6400-29f0-41da-9c22-0443d88dca9d\") " Oct 10 16:17:53 crc kubenswrapper[4788]: I1010 16:17:53.423469 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2cd6400-29f0-41da-9c22-0443d88dca9d-kube-api-access-qrzj2" (OuterVolumeSpecName: "kube-api-access-qrzj2") pod "f2cd6400-29f0-41da-9c22-0443d88dca9d" (UID: "f2cd6400-29f0-41da-9c22-0443d88dca9d"). InnerVolumeSpecName "kube-api-access-qrzj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:17:53 crc kubenswrapper[4788]: I1010 16:17:53.519405 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrzj2\" (UniqueName: \"kubernetes.io/projected/f2cd6400-29f0-41da-9c22-0443d88dca9d-kube-api-access-qrzj2\") on node \"crc\" DevicePath \"\"" Oct 10 16:17:54 crc kubenswrapper[4788]: I1010 16:17:54.049056 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c7cb-account-create-v5t9s" event={"ID":"f2cd6400-29f0-41da-9c22-0443d88dca9d","Type":"ContainerDied","Data":"be7fe79464afa39c1c85a2261047c9cc8df9a471a9381cdcfb11fa7f8d57048f"} Oct 10 16:17:54 crc kubenswrapper[4788]: I1010 16:17:54.049448 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be7fe79464afa39c1c85a2261047c9cc8df9a471a9381cdcfb11fa7f8d57048f" Oct 10 16:17:54 crc kubenswrapper[4788]: I1010 16:17:54.049257 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c7cb-account-create-v5t9s" Oct 10 16:17:55 crc kubenswrapper[4788]: I1010 16:17:55.958692 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-d2bgn"] Oct 10 16:17:55 crc kubenswrapper[4788]: E1010 16:17:55.959077 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cd6400-29f0-41da-9c22-0443d88dca9d" containerName="mariadb-account-create" Oct 10 16:17:55 crc kubenswrapper[4788]: I1010 16:17:55.959088 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cd6400-29f0-41da-9c22-0443d88dca9d" containerName="mariadb-account-create" Oct 10 16:17:55 crc kubenswrapper[4788]: I1010 16:17:55.959274 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2cd6400-29f0-41da-9c22-0443d88dca9d" containerName="mariadb-account-create" Oct 10 16:17:55 crc kubenswrapper[4788]: I1010 16:17:55.959829 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:55 crc kubenswrapper[4788]: I1010 16:17:55.961877 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 10 16:17:55 crc kubenswrapper[4788]: I1010 16:17:55.962139 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5twzb" Oct 10 16:17:55 crc kubenswrapper[4788]: I1010 16:17:55.962571 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 10 16:17:55 crc kubenswrapper[4788]: I1010 16:17:55.974593 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-d2bgn"] Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.011732 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bb448fdbf-sbkd8"] Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.013291 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.030761 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bb448fdbf-sbkd8"] Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.093004 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/495332ff-4fbc-44db-bb18-bd59715f8e4e-logs\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.093421 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-scripts\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.093492 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-combined-ca-bundle\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.093521 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-config-data\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.093650 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgfj8\" (UniqueName: \"kubernetes.io/projected/495332ff-4fbc-44db-bb18-bd59715f8e4e-kube-api-access-zgfj8\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.195954 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-config\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196012 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-dns-svc\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196052 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-scripts\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196105 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-combined-ca-bundle\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196124 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-config-data\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196213 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgfj8\" (UniqueName: \"kubernetes.io/projected/495332ff-4fbc-44db-bb18-bd59715f8e4e-kube-api-access-zgfj8\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196271 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkmh2\" (UniqueName: \"kubernetes.io/projected/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-kube-api-access-bkmh2\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196305 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-nb\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196325 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-sb\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196358 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/495332ff-4fbc-44db-bb18-bd59715f8e4e-logs\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.196963 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/495332ff-4fbc-44db-bb18-bd59715f8e4e-logs\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.204519 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-config-data\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.206533 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-scripts\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.207953 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-combined-ca-bundle\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.218531 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgfj8\" (UniqueName: \"kubernetes.io/projected/495332ff-4fbc-44db-bb18-bd59715f8e4e-kube-api-access-zgfj8\") pod \"placement-db-sync-d2bgn\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.288189 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-d2bgn" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.298273 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkmh2\" (UniqueName: \"kubernetes.io/projected/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-kube-api-access-bkmh2\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.298339 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-nb\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.298368 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-sb\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.298418 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-config\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.298449 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-dns-svc\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.299643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-dns-svc\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.301663 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-config\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.302201 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-sb\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.302627 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-nb\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.321360 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkmh2\" (UniqueName: \"kubernetes.io/projected/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-kube-api-access-bkmh2\") pod \"dnsmasq-dns-bb448fdbf-sbkd8\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.337939 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.821945 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-d2bgn"] Oct 10 16:17:56 crc kubenswrapper[4788]: I1010 16:17:56.903738 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bb448fdbf-sbkd8"] Oct 10 16:17:57 crc kubenswrapper[4788]: I1010 16:17:57.080672 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-d2bgn" event={"ID":"495332ff-4fbc-44db-bb18-bd59715f8e4e","Type":"ContainerStarted","Data":"516a629af29b70aa5c5fb6a8ad3bb4fa3a67c404433276b525663a3b71197d1f"} Oct 10 16:17:57 crc kubenswrapper[4788]: I1010 16:17:57.081127 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-d2bgn" event={"ID":"495332ff-4fbc-44db-bb18-bd59715f8e4e","Type":"ContainerStarted","Data":"fff5cce39f998d57db2567dba748c1bf79e454f0952d9f922b8c05f157aeb3f6"} Oct 10 16:17:57 crc kubenswrapper[4788]: I1010 16:17:57.081819 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" event={"ID":"8ea194fc-2dde-490a-b9d5-6f17dfd3af19","Type":"ContainerStarted","Data":"4e12fbe4a231c17d546ffdb2edbc42eafa2fe698d4cd7e3b78129a99ffb76cdb"} Oct 10 16:17:57 crc kubenswrapper[4788]: I1010 16:17:57.112546 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-d2bgn" podStartSLOduration=2.112525455 podStartE2EDuration="2.112525455s" podCreationTimestamp="2025-10-10 16:17:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:17:57.101536023 +0000 UTC m=+5579.551251571" watchObservedRunningTime="2025-10-10 16:17:57.112525455 +0000 UTC m=+5579.562241003" Oct 10 16:17:58 crc kubenswrapper[4788]: I1010 16:17:58.096434 4788 generic.go:334] "Generic (PLEG): container finished" podID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" containerID="ace6045c4c7d1dc4d05567790fea276f0de9572dd348e214200222f24131998d" exitCode=0 Oct 10 16:17:58 crc kubenswrapper[4788]: I1010 16:17:58.096556 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" event={"ID":"8ea194fc-2dde-490a-b9d5-6f17dfd3af19","Type":"ContainerDied","Data":"ace6045c4c7d1dc4d05567790fea276f0de9572dd348e214200222f24131998d"} Oct 10 16:17:59 crc kubenswrapper[4788]: I1010 16:17:59.109301 4788 generic.go:334] "Generic (PLEG): container finished" podID="495332ff-4fbc-44db-bb18-bd59715f8e4e" containerID="516a629af29b70aa5c5fb6a8ad3bb4fa3a67c404433276b525663a3b71197d1f" exitCode=0 Oct 10 16:17:59 crc kubenswrapper[4788]: I1010 16:17:59.109380 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-d2bgn" event={"ID":"495332ff-4fbc-44db-bb18-bd59715f8e4e","Type":"ContainerDied","Data":"516a629af29b70aa5c5fb6a8ad3bb4fa3a67c404433276b525663a3b71197d1f"} Oct 10 16:17:59 crc kubenswrapper[4788]: I1010 16:17:59.113624 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" event={"ID":"8ea194fc-2dde-490a-b9d5-6f17dfd3af19","Type":"ContainerStarted","Data":"426a1f0c09305f48ce78f7012b0c38962132fd4e6911d21378d09517cf78cf68"} Oct 10 16:17:59 crc kubenswrapper[4788]: I1010 16:17:59.113937 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:17:59 crc kubenswrapper[4788]: I1010 16:17:59.159578 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" podStartSLOduration=4.159552063 podStartE2EDuration="4.159552063s" podCreationTimestamp="2025-10-10 16:17:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:17:59.151017433 +0000 UTC m=+5581.600733001" watchObservedRunningTime="2025-10-10 16:17:59.159552063 +0000 UTC m=+5581.609267641" Oct 10 16:17:59 crc kubenswrapper[4788]: I1010 16:17:59.406047 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:17:59 crc kubenswrapper[4788]: I1010 16:17:59.406425 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.522022 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-d2bgn" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.687883 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/495332ff-4fbc-44db-bb18-bd59715f8e4e-logs\") pod \"495332ff-4fbc-44db-bb18-bd59715f8e4e\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.688110 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgfj8\" (UniqueName: \"kubernetes.io/projected/495332ff-4fbc-44db-bb18-bd59715f8e4e-kube-api-access-zgfj8\") pod \"495332ff-4fbc-44db-bb18-bd59715f8e4e\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.688234 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-combined-ca-bundle\") pod \"495332ff-4fbc-44db-bb18-bd59715f8e4e\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.688269 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-config-data\") pod \"495332ff-4fbc-44db-bb18-bd59715f8e4e\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.688320 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-scripts\") pod \"495332ff-4fbc-44db-bb18-bd59715f8e4e\" (UID: \"495332ff-4fbc-44db-bb18-bd59715f8e4e\") " Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.688558 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/495332ff-4fbc-44db-bb18-bd59715f8e4e-logs" (OuterVolumeSpecName: "logs") pod "495332ff-4fbc-44db-bb18-bd59715f8e4e" (UID: "495332ff-4fbc-44db-bb18-bd59715f8e4e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.689114 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/495332ff-4fbc-44db-bb18-bd59715f8e4e-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.693955 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-scripts" (OuterVolumeSpecName: "scripts") pod "495332ff-4fbc-44db-bb18-bd59715f8e4e" (UID: "495332ff-4fbc-44db-bb18-bd59715f8e4e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.694510 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/495332ff-4fbc-44db-bb18-bd59715f8e4e-kube-api-access-zgfj8" (OuterVolumeSpecName: "kube-api-access-zgfj8") pod "495332ff-4fbc-44db-bb18-bd59715f8e4e" (UID: "495332ff-4fbc-44db-bb18-bd59715f8e4e"). InnerVolumeSpecName "kube-api-access-zgfj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.722531 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "495332ff-4fbc-44db-bb18-bd59715f8e4e" (UID: "495332ff-4fbc-44db-bb18-bd59715f8e4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.741286 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-config-data" (OuterVolumeSpecName: "config-data") pod "495332ff-4fbc-44db-bb18-bd59715f8e4e" (UID: "495332ff-4fbc-44db-bb18-bd59715f8e4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.791537 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgfj8\" (UniqueName: \"kubernetes.io/projected/495332ff-4fbc-44db-bb18-bd59715f8e4e-kube-api-access-zgfj8\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.791577 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.791591 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:00 crc kubenswrapper[4788]: I1010 16:18:00.791607 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/495332ff-4fbc-44db-bb18-bd59715f8e4e-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.134767 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-d2bgn" event={"ID":"495332ff-4fbc-44db-bb18-bd59715f8e4e","Type":"ContainerDied","Data":"fff5cce39f998d57db2567dba748c1bf79e454f0952d9f922b8c05f157aeb3f6"} Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.134799 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-d2bgn" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.134815 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fff5cce39f998d57db2567dba748c1bf79e454f0952d9f922b8c05f157aeb3f6" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.327082 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-75f45f9948-dj9bl"] Oct 10 16:18:01 crc kubenswrapper[4788]: E1010 16:18:01.327554 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495332ff-4fbc-44db-bb18-bd59715f8e4e" containerName="placement-db-sync" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.327580 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="495332ff-4fbc-44db-bb18-bd59715f8e4e" containerName="placement-db-sync" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.327825 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="495332ff-4fbc-44db-bb18-bd59715f8e4e" containerName="placement-db-sync" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.328995 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.333316 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.333646 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5twzb" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.334001 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.343131 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75f45f9948-dj9bl"] Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.504404 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5tgk\" (UniqueName: \"kubernetes.io/projected/cade6027-5453-4492-919e-62a481d01439-kube-api-access-c5tgk\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.504822 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cade6027-5453-4492-919e-62a481d01439-logs\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.504977 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-combined-ca-bundle\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.505108 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-scripts\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.505254 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-config-data\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.607574 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cade6027-5453-4492-919e-62a481d01439-logs\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.607669 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-combined-ca-bundle\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.607711 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-scripts\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.607754 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-config-data\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.607814 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5tgk\" (UniqueName: \"kubernetes.io/projected/cade6027-5453-4492-919e-62a481d01439-kube-api-access-c5tgk\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.608574 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cade6027-5453-4492-919e-62a481d01439-logs\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.615037 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-scripts\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.615502 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-config-data\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.618742 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cade6027-5453-4492-919e-62a481d01439-combined-ca-bundle\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.626448 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5tgk\" (UniqueName: \"kubernetes.io/projected/cade6027-5453-4492-919e-62a481d01439-kube-api-access-c5tgk\") pod \"placement-75f45f9948-dj9bl\" (UID: \"cade6027-5453-4492-919e-62a481d01439\") " pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:01 crc kubenswrapper[4788]: I1010 16:18:01.654379 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:02 crc kubenswrapper[4788]: I1010 16:18:02.078309 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75f45f9948-dj9bl"] Oct 10 16:18:02 crc kubenswrapper[4788]: W1010 16:18:02.088055 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcade6027_5453_4492_919e_62a481d01439.slice/crio-5d399962c16fe2fd2bf9f247d660e26ad3c3b548ceb87f73cc2af1383c45e1f3 WatchSource:0}: Error finding container 5d399962c16fe2fd2bf9f247d660e26ad3c3b548ceb87f73cc2af1383c45e1f3: Status 404 returned error can't find the container with id 5d399962c16fe2fd2bf9f247d660e26ad3c3b548ceb87f73cc2af1383c45e1f3 Oct 10 16:18:02 crc kubenswrapper[4788]: I1010 16:18:02.148358 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75f45f9948-dj9bl" event={"ID":"cade6027-5453-4492-919e-62a481d01439","Type":"ContainerStarted","Data":"5d399962c16fe2fd2bf9f247d660e26ad3c3b548ceb87f73cc2af1383c45e1f3"} Oct 10 16:18:03 crc kubenswrapper[4788]: I1010 16:18:03.160630 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75f45f9948-dj9bl" event={"ID":"cade6027-5453-4492-919e-62a481d01439","Type":"ContainerStarted","Data":"7c7379b778f362e7a30fafe30aca49dd1e411bf0e979518d1a42867d80157ae8"} Oct 10 16:18:03 crc kubenswrapper[4788]: I1010 16:18:03.160701 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75f45f9948-dj9bl" event={"ID":"cade6027-5453-4492-919e-62a481d01439","Type":"ContainerStarted","Data":"8acf9e896854565fd0eb84dc62dd7ae5a3b64a80fc1621ef26f08cd740a0333b"} Oct 10 16:18:03 crc kubenswrapper[4788]: I1010 16:18:03.162156 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:03 crc kubenswrapper[4788]: I1010 16:18:03.185714 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-75f45f9948-dj9bl" podStartSLOduration=2.185694172 podStartE2EDuration="2.185694172s" podCreationTimestamp="2025-10-10 16:18:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:18:03.175991763 +0000 UTC m=+5585.625707341" watchObservedRunningTime="2025-10-10 16:18:03.185694172 +0000 UTC m=+5585.635409730" Oct 10 16:18:04 crc kubenswrapper[4788]: I1010 16:18:04.170407 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.340383 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.395962 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f8774dc7-5jnft"] Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.396305 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" podUID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" containerName="dnsmasq-dns" containerID="cri-o://85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41" gracePeriod=10 Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.818357 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.916960 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-config\") pod \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.917093 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7gbm\" (UniqueName: \"kubernetes.io/projected/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-kube-api-access-r7gbm\") pod \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.917191 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-dns-svc\") pod \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.917284 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-sb\") pod \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.917302 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-nb\") pod \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\" (UID: \"8e9df50e-c08f-4bf7-ad94-22813cc9efa6\") " Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.923435 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-kube-api-access-r7gbm" (OuterVolumeSpecName: "kube-api-access-r7gbm") pod "8e9df50e-c08f-4bf7-ad94-22813cc9efa6" (UID: "8e9df50e-c08f-4bf7-ad94-22813cc9efa6"). InnerVolumeSpecName "kube-api-access-r7gbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.960030 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-config" (OuterVolumeSpecName: "config") pod "8e9df50e-c08f-4bf7-ad94-22813cc9efa6" (UID: "8e9df50e-c08f-4bf7-ad94-22813cc9efa6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.961210 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e9df50e-c08f-4bf7-ad94-22813cc9efa6" (UID: "8e9df50e-c08f-4bf7-ad94-22813cc9efa6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.962363 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e9df50e-c08f-4bf7-ad94-22813cc9efa6" (UID: "8e9df50e-c08f-4bf7-ad94-22813cc9efa6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:18:06 crc kubenswrapper[4788]: I1010 16:18:06.963241 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e9df50e-c08f-4bf7-ad94-22813cc9efa6" (UID: "8e9df50e-c08f-4bf7-ad94-22813cc9efa6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.020529 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.020585 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.020598 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.020610 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7gbm\" (UniqueName: \"kubernetes.io/projected/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-kube-api-access-r7gbm\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.020643 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e9df50e-c08f-4bf7-ad94-22813cc9efa6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.201692 4788 generic.go:334] "Generic (PLEG): container finished" podID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" containerID="85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41" exitCode=0 Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.201744 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.201743 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" event={"ID":"8e9df50e-c08f-4bf7-ad94-22813cc9efa6","Type":"ContainerDied","Data":"85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41"} Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.201835 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f8774dc7-5jnft" event={"ID":"8e9df50e-c08f-4bf7-ad94-22813cc9efa6","Type":"ContainerDied","Data":"f2e969910fc76eb2320e1348ca4178d8f5ebeb3ab4055f0d63a6a8c3ec059caf"} Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.201876 4788 scope.go:117] "RemoveContainer" containerID="85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.226508 4788 scope.go:117] "RemoveContainer" containerID="bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.237087 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f8774dc7-5jnft"] Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.245063 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f8774dc7-5jnft"] Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.265994 4788 scope.go:117] "RemoveContainer" containerID="85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41" Oct 10 16:18:07 crc kubenswrapper[4788]: E1010 16:18:07.266474 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41\": container with ID starting with 85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41 not found: ID does not exist" containerID="85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.266505 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41"} err="failed to get container status \"85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41\": rpc error: code = NotFound desc = could not find container \"85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41\": container with ID starting with 85fe7f54828a2dec72eae1a6e00a2c84d73c78d0dabaa1d3fa244c50675eea41 not found: ID does not exist" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.266526 4788 scope.go:117] "RemoveContainer" containerID="bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b" Oct 10 16:18:07 crc kubenswrapper[4788]: E1010 16:18:07.266871 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b\": container with ID starting with bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b not found: ID does not exist" containerID="bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b" Oct 10 16:18:07 crc kubenswrapper[4788]: I1010 16:18:07.266929 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b"} err="failed to get container status \"bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b\": rpc error: code = NotFound desc = could not find container \"bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b\": container with ID starting with bd09a47fc815757d3863edbec3878fd6638011069cb737f7321981ca9ad4118b not found: ID does not exist" Oct 10 16:18:08 crc kubenswrapper[4788]: I1010 16:18:08.246878 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" path="/var/lib/kubelet/pods/8e9df50e-c08f-4bf7-ad94-22813cc9efa6/volumes" Oct 10 16:18:29 crc kubenswrapper[4788]: I1010 16:18:29.407790 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:18:29 crc kubenswrapper[4788]: I1010 16:18:29.409265 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:18:29 crc kubenswrapper[4788]: I1010 16:18:29.409336 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:18:29 crc kubenswrapper[4788]: I1010 16:18:29.409981 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:18:29 crc kubenswrapper[4788]: I1010 16:18:29.410043 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" gracePeriod=600 Oct 10 16:18:29 crc kubenswrapper[4788]: E1010 16:18:29.532458 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:18:30 crc kubenswrapper[4788]: I1010 16:18:30.448193 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" exitCode=0 Oct 10 16:18:30 crc kubenswrapper[4788]: I1010 16:18:30.448270 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2"} Oct 10 16:18:30 crc kubenswrapper[4788]: I1010 16:18:30.448507 4788 scope.go:117] "RemoveContainer" containerID="f2a936fbf95226f44505aa4c234e52e71830372cdc6c6f705ee31bf2d04b44c1" Oct 10 16:18:30 crc kubenswrapper[4788]: I1010 16:18:30.448970 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:18:30 crc kubenswrapper[4788]: E1010 16:18:30.449212 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:18:32 crc kubenswrapper[4788]: I1010 16:18:32.613171 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:32 crc kubenswrapper[4788]: I1010 16:18:32.620281 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-75f45f9948-dj9bl" Oct 10 16:18:42 crc kubenswrapper[4788]: I1010 16:18:42.234134 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:18:42 crc kubenswrapper[4788]: E1010 16:18:42.234880 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.769225 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4x45l"] Oct 10 16:18:55 crc kubenswrapper[4788]: E1010 16:18:55.770114 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" containerName="dnsmasq-dns" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.770128 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" containerName="dnsmasq-dns" Oct 10 16:18:55 crc kubenswrapper[4788]: E1010 16:18:55.770174 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" containerName="init" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.770182 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" containerName="init" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.770347 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e9df50e-c08f-4bf7-ad94-22813cc9efa6" containerName="dnsmasq-dns" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.771018 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4x45l" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.787454 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4x45l"] Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.815384 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvrhf\" (UniqueName: \"kubernetes.io/projected/7f418231-e326-423e-948e-fa69f7a727e0-kube-api-access-qvrhf\") pod \"nova-api-db-create-4x45l\" (UID: \"7f418231-e326-423e-948e-fa69f7a727e0\") " pod="openstack/nova-api-db-create-4x45l" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.862773 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jn9s4"] Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.863841 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jn9s4" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.874030 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jn9s4"] Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.917215 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvrhf\" (UniqueName: \"kubernetes.io/projected/7f418231-e326-423e-948e-fa69f7a727e0-kube-api-access-qvrhf\") pod \"nova-api-db-create-4x45l\" (UID: \"7f418231-e326-423e-948e-fa69f7a727e0\") " pod="openstack/nova-api-db-create-4x45l" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.917339 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mjcm\" (UniqueName: \"kubernetes.io/projected/77a38426-b4b3-4576-99de-b430003962d5-kube-api-access-5mjcm\") pod \"nova-cell0-db-create-jn9s4\" (UID: \"77a38426-b4b3-4576-99de-b430003962d5\") " pod="openstack/nova-cell0-db-create-jn9s4" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.938776 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvrhf\" (UniqueName: \"kubernetes.io/projected/7f418231-e326-423e-948e-fa69f7a727e0-kube-api-access-qvrhf\") pod \"nova-api-db-create-4x45l\" (UID: \"7f418231-e326-423e-948e-fa69f7a727e0\") " pod="openstack/nova-api-db-create-4x45l" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.962810 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-nmcdb"] Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.963951 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nmcdb" Oct 10 16:18:55 crc kubenswrapper[4788]: I1010 16:18:55.983187 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-nmcdb"] Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.021314 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mjcm\" (UniqueName: \"kubernetes.io/projected/77a38426-b4b3-4576-99de-b430003962d5-kube-api-access-5mjcm\") pod \"nova-cell0-db-create-jn9s4\" (UID: \"77a38426-b4b3-4576-99de-b430003962d5\") " pod="openstack/nova-cell0-db-create-jn9s4" Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.021406 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-796sg\" (UniqueName: \"kubernetes.io/projected/80be9ca4-09b7-4907-ac98-1423d0ffa711-kube-api-access-796sg\") pod \"nova-cell1-db-create-nmcdb\" (UID: \"80be9ca4-09b7-4907-ac98-1423d0ffa711\") " pod="openstack/nova-cell1-db-create-nmcdb" Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.038340 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mjcm\" (UniqueName: \"kubernetes.io/projected/77a38426-b4b3-4576-99de-b430003962d5-kube-api-access-5mjcm\") pod \"nova-cell0-db-create-jn9s4\" (UID: \"77a38426-b4b3-4576-99de-b430003962d5\") " pod="openstack/nova-cell0-db-create-jn9s4" Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.093360 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4x45l" Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.124233 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-796sg\" (UniqueName: \"kubernetes.io/projected/80be9ca4-09b7-4907-ac98-1423d0ffa711-kube-api-access-796sg\") pod \"nova-cell1-db-create-nmcdb\" (UID: \"80be9ca4-09b7-4907-ac98-1423d0ffa711\") " pod="openstack/nova-cell1-db-create-nmcdb" Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.144244 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-796sg\" (UniqueName: \"kubernetes.io/projected/80be9ca4-09b7-4907-ac98-1423d0ffa711-kube-api-access-796sg\") pod \"nova-cell1-db-create-nmcdb\" (UID: \"80be9ca4-09b7-4907-ac98-1423d0ffa711\") " pod="openstack/nova-cell1-db-create-nmcdb" Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.188617 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jn9s4" Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.239898 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:18:56 crc kubenswrapper[4788]: E1010 16:18:56.240122 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:18:56 crc kubenswrapper[4788]: I1010 16:18:56.302952 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nmcdb" Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.158043 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jn9s4"] Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.172992 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4x45l"] Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.174927 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-nmcdb"] Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.699253 4788 generic.go:334] "Generic (PLEG): container finished" podID="77a38426-b4b3-4576-99de-b430003962d5" containerID="c4692339c259979db1616d69dbb17890301f186aa97725329c4e309b4e3fad24" exitCode=0 Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.699347 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jn9s4" event={"ID":"77a38426-b4b3-4576-99de-b430003962d5","Type":"ContainerDied","Data":"c4692339c259979db1616d69dbb17890301f186aa97725329c4e309b4e3fad24"} Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.699596 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jn9s4" event={"ID":"77a38426-b4b3-4576-99de-b430003962d5","Type":"ContainerStarted","Data":"b3851589311c057138c835eb88a100bdc8d153aa4fbb37d17d88a4f961220884"} Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.701473 4788 generic.go:334] "Generic (PLEG): container finished" podID="80be9ca4-09b7-4907-ac98-1423d0ffa711" containerID="09e234a86a10f7cbcf7659080fecb1c1a6137b12d15dd15790db7f8d42b12708" exitCode=0 Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.701535 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nmcdb" event={"ID":"80be9ca4-09b7-4907-ac98-1423d0ffa711","Type":"ContainerDied","Data":"09e234a86a10f7cbcf7659080fecb1c1a6137b12d15dd15790db7f8d42b12708"} Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.701558 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nmcdb" event={"ID":"80be9ca4-09b7-4907-ac98-1423d0ffa711","Type":"ContainerStarted","Data":"e75be7bb9479429bc9565136f2a5ceebb1dc7e9a9967afd312d865e7818b5e12"} Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.703044 4788 generic.go:334] "Generic (PLEG): container finished" podID="7f418231-e326-423e-948e-fa69f7a727e0" containerID="bd2e4d573e6fa9ca14b21d675411bc87317c1059fa7b0a0c9dec5b6c7ad38a68" exitCode=0 Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.703072 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4x45l" event={"ID":"7f418231-e326-423e-948e-fa69f7a727e0","Type":"ContainerDied","Data":"bd2e4d573e6fa9ca14b21d675411bc87317c1059fa7b0a0c9dec5b6c7ad38a68"} Oct 10 16:18:57 crc kubenswrapper[4788]: I1010 16:18:57.703086 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4x45l" event={"ID":"7f418231-e326-423e-948e-fa69f7a727e0","Type":"ContainerStarted","Data":"7305ed1723c5d3174ce1b654eef3302bcb6c3f301c40b2739ab9a08aa5743835"} Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.151890 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4x45l" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.159163 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jn9s4" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.171257 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nmcdb" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.296590 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvrhf\" (UniqueName: \"kubernetes.io/projected/7f418231-e326-423e-948e-fa69f7a727e0-kube-api-access-qvrhf\") pod \"7f418231-e326-423e-948e-fa69f7a727e0\" (UID: \"7f418231-e326-423e-948e-fa69f7a727e0\") " Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.296772 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mjcm\" (UniqueName: \"kubernetes.io/projected/77a38426-b4b3-4576-99de-b430003962d5-kube-api-access-5mjcm\") pod \"77a38426-b4b3-4576-99de-b430003962d5\" (UID: \"77a38426-b4b3-4576-99de-b430003962d5\") " Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.296796 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-796sg\" (UniqueName: \"kubernetes.io/projected/80be9ca4-09b7-4907-ac98-1423d0ffa711-kube-api-access-796sg\") pod \"80be9ca4-09b7-4907-ac98-1423d0ffa711\" (UID: \"80be9ca4-09b7-4907-ac98-1423d0ffa711\") " Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.304176 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f418231-e326-423e-948e-fa69f7a727e0-kube-api-access-qvrhf" (OuterVolumeSpecName: "kube-api-access-qvrhf") pod "7f418231-e326-423e-948e-fa69f7a727e0" (UID: "7f418231-e326-423e-948e-fa69f7a727e0"). InnerVolumeSpecName "kube-api-access-qvrhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.304352 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80be9ca4-09b7-4907-ac98-1423d0ffa711-kube-api-access-796sg" (OuterVolumeSpecName: "kube-api-access-796sg") pod "80be9ca4-09b7-4907-ac98-1423d0ffa711" (UID: "80be9ca4-09b7-4907-ac98-1423d0ffa711"). InnerVolumeSpecName "kube-api-access-796sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.305743 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77a38426-b4b3-4576-99de-b430003962d5-kube-api-access-5mjcm" (OuterVolumeSpecName: "kube-api-access-5mjcm") pod "77a38426-b4b3-4576-99de-b430003962d5" (UID: "77a38426-b4b3-4576-99de-b430003962d5"). InnerVolumeSpecName "kube-api-access-5mjcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.399530 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-796sg\" (UniqueName: \"kubernetes.io/projected/80be9ca4-09b7-4907-ac98-1423d0ffa711-kube-api-access-796sg\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.400023 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvrhf\" (UniqueName: \"kubernetes.io/projected/7f418231-e326-423e-948e-fa69f7a727e0-kube-api-access-qvrhf\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.400039 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mjcm\" (UniqueName: \"kubernetes.io/projected/77a38426-b4b3-4576-99de-b430003962d5-kube-api-access-5mjcm\") on node \"crc\" DevicePath \"\"" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.738649 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4x45l" event={"ID":"7f418231-e326-423e-948e-fa69f7a727e0","Type":"ContainerDied","Data":"7305ed1723c5d3174ce1b654eef3302bcb6c3f301c40b2739ab9a08aa5743835"} Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.738692 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7305ed1723c5d3174ce1b654eef3302bcb6c3f301c40b2739ab9a08aa5743835" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.738693 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4x45l" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.740538 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jn9s4" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.740556 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jn9s4" event={"ID":"77a38426-b4b3-4576-99de-b430003962d5","Type":"ContainerDied","Data":"b3851589311c057138c835eb88a100bdc8d153aa4fbb37d17d88a4f961220884"} Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.740605 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3851589311c057138c835eb88a100bdc8d153aa4fbb37d17d88a4f961220884" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.743363 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-nmcdb" event={"ID":"80be9ca4-09b7-4907-ac98-1423d0ffa711","Type":"ContainerDied","Data":"e75be7bb9479429bc9565136f2a5ceebb1dc7e9a9967afd312d865e7818b5e12"} Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.743402 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e75be7bb9479429bc9565136f2a5ceebb1dc7e9a9967afd312d865e7818b5e12" Oct 10 16:18:59 crc kubenswrapper[4788]: I1010 16:18:59.743436 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-nmcdb" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.005869 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c927-account-create-vwdhf"] Oct 10 16:19:06 crc kubenswrapper[4788]: E1010 16:19:06.007117 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80be9ca4-09b7-4907-ac98-1423d0ffa711" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.007163 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="80be9ca4-09b7-4907-ac98-1423d0ffa711" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: E1010 16:19:06.007186 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f418231-e326-423e-948e-fa69f7a727e0" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.007198 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f418231-e326-423e-948e-fa69f7a727e0" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: E1010 16:19:06.007219 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a38426-b4b3-4576-99de-b430003962d5" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.007232 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a38426-b4b3-4576-99de-b430003962d5" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.007526 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f418231-e326-423e-948e-fa69f7a727e0" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.007559 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="80be9ca4-09b7-4907-ac98-1423d0ffa711" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.007590 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="77a38426-b4b3-4576-99de-b430003962d5" containerName="mariadb-database-create" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.008814 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c927-account-create-vwdhf" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.012511 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.016255 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c927-account-create-vwdhf"] Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.143109 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jksbv\" (UniqueName: \"kubernetes.io/projected/7cf0be59-5a06-45c5-b913-d75c1e4096c0-kube-api-access-jksbv\") pod \"nova-api-c927-account-create-vwdhf\" (UID: \"7cf0be59-5a06-45c5-b913-d75c1e4096c0\") " pod="openstack/nova-api-c927-account-create-vwdhf" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.188480 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0d75-account-create-kb2xh"] Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.189578 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0d75-account-create-kb2xh" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.194892 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.202889 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0d75-account-create-kb2xh"] Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.244735 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jksbv\" (UniqueName: \"kubernetes.io/projected/7cf0be59-5a06-45c5-b913-d75c1e4096c0-kube-api-access-jksbv\") pod \"nova-api-c927-account-create-vwdhf\" (UID: \"7cf0be59-5a06-45c5-b913-d75c1e4096c0\") " pod="openstack/nova-api-c927-account-create-vwdhf" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.265426 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jksbv\" (UniqueName: \"kubernetes.io/projected/7cf0be59-5a06-45c5-b913-d75c1e4096c0-kube-api-access-jksbv\") pod \"nova-api-c927-account-create-vwdhf\" (UID: \"7cf0be59-5a06-45c5-b913-d75c1e4096c0\") " pod="openstack/nova-api-c927-account-create-vwdhf" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.339097 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c927-account-create-vwdhf" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.346092 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6z9j\" (UniqueName: \"kubernetes.io/projected/a7d44061-2a00-4a40-94c7-0c56e37d55c9-kube-api-access-l6z9j\") pod \"nova-cell0-0d75-account-create-kb2xh\" (UID: \"a7d44061-2a00-4a40-94c7-0c56e37d55c9\") " pod="openstack/nova-cell0-0d75-account-create-kb2xh" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.404037 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8011-account-create-hb7s5"] Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.405272 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8011-account-create-hb7s5" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.408322 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.412699 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8011-account-create-hb7s5"] Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.448056 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6z9j\" (UniqueName: \"kubernetes.io/projected/a7d44061-2a00-4a40-94c7-0c56e37d55c9-kube-api-access-l6z9j\") pod \"nova-cell0-0d75-account-create-kb2xh\" (UID: \"a7d44061-2a00-4a40-94c7-0c56e37d55c9\") " pod="openstack/nova-cell0-0d75-account-create-kb2xh" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.466964 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6z9j\" (UniqueName: \"kubernetes.io/projected/a7d44061-2a00-4a40-94c7-0c56e37d55c9-kube-api-access-l6z9j\") pod \"nova-cell0-0d75-account-create-kb2xh\" (UID: \"a7d44061-2a00-4a40-94c7-0c56e37d55c9\") " pod="openstack/nova-cell0-0d75-account-create-kb2xh" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.513841 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0d75-account-create-kb2xh" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.550679 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4w75\" (UniqueName: \"kubernetes.io/projected/1655641f-54ec-48c1-bbb6-21e801e6a54f-kube-api-access-v4w75\") pod \"nova-cell1-8011-account-create-hb7s5\" (UID: \"1655641f-54ec-48c1-bbb6-21e801e6a54f\") " pod="openstack/nova-cell1-8011-account-create-hb7s5" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.652635 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4w75\" (UniqueName: \"kubernetes.io/projected/1655641f-54ec-48c1-bbb6-21e801e6a54f-kube-api-access-v4w75\") pod \"nova-cell1-8011-account-create-hb7s5\" (UID: \"1655641f-54ec-48c1-bbb6-21e801e6a54f\") " pod="openstack/nova-cell1-8011-account-create-hb7s5" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.668261 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4w75\" (UniqueName: \"kubernetes.io/projected/1655641f-54ec-48c1-bbb6-21e801e6a54f-kube-api-access-v4w75\") pod \"nova-cell1-8011-account-create-hb7s5\" (UID: \"1655641f-54ec-48c1-bbb6-21e801e6a54f\") " pod="openstack/nova-cell1-8011-account-create-hb7s5" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.761375 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8011-account-create-hb7s5" Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.776762 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c927-account-create-vwdhf"] Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.813387 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c927-account-create-vwdhf" event={"ID":"7cf0be59-5a06-45c5-b913-d75c1e4096c0","Type":"ContainerStarted","Data":"69f0a676e849637f6e898fe0d903c380d6115091913dd03c0eb8081c0bba151a"} Oct 10 16:19:06 crc kubenswrapper[4788]: I1010 16:19:06.928198 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0d75-account-create-kb2xh"] Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.190072 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8011-account-create-hb7s5"] Oct 10 16:19:07 crc kubenswrapper[4788]: W1010 16:19:07.223314 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1655641f_54ec_48c1_bbb6_21e801e6a54f.slice/crio-85e356354f9b0cdddf2dab4d7c008dc38900266d074fdf8ef5adaaefdb621723 WatchSource:0}: Error finding container 85e356354f9b0cdddf2dab4d7c008dc38900266d074fdf8ef5adaaefdb621723: Status 404 returned error can't find the container with id 85e356354f9b0cdddf2dab4d7c008dc38900266d074fdf8ef5adaaefdb621723 Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.233276 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:19:07 crc kubenswrapper[4788]: E1010 16:19:07.233504 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.824280 4788 generic.go:334] "Generic (PLEG): container finished" podID="7cf0be59-5a06-45c5-b913-d75c1e4096c0" containerID="261022933c054799e06153249db6b73508f0a918af13e3afcbf4fdb0544deb67" exitCode=0 Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.824357 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c927-account-create-vwdhf" event={"ID":"7cf0be59-5a06-45c5-b913-d75c1e4096c0","Type":"ContainerDied","Data":"261022933c054799e06153249db6b73508f0a918af13e3afcbf4fdb0544deb67"} Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.828966 4788 generic.go:334] "Generic (PLEG): container finished" podID="1655641f-54ec-48c1-bbb6-21e801e6a54f" containerID="99ae820df5735d54c6b0124c9827eac416b9507d6069078f1b4cd7d6a1b1e424" exitCode=0 Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.829317 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8011-account-create-hb7s5" event={"ID":"1655641f-54ec-48c1-bbb6-21e801e6a54f","Type":"ContainerDied","Data":"99ae820df5735d54c6b0124c9827eac416b9507d6069078f1b4cd7d6a1b1e424"} Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.830464 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8011-account-create-hb7s5" event={"ID":"1655641f-54ec-48c1-bbb6-21e801e6a54f","Type":"ContainerStarted","Data":"85e356354f9b0cdddf2dab4d7c008dc38900266d074fdf8ef5adaaefdb621723"} Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.833336 4788 generic.go:334] "Generic (PLEG): container finished" podID="a7d44061-2a00-4a40-94c7-0c56e37d55c9" containerID="2ff29cbbeaef5875febfd11818203b781590c9a4d5ee1b53f60df32688f3cd5a" exitCode=0 Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.833374 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0d75-account-create-kb2xh" event={"ID":"a7d44061-2a00-4a40-94c7-0c56e37d55c9","Type":"ContainerDied","Data":"2ff29cbbeaef5875febfd11818203b781590c9a4d5ee1b53f60df32688f3cd5a"} Oct 10 16:19:07 crc kubenswrapper[4788]: I1010 16:19:07.833394 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0d75-account-create-kb2xh" event={"ID":"a7d44061-2a00-4a40-94c7-0c56e37d55c9","Type":"ContainerStarted","Data":"88f36a385e300d49eff13c35709907f2489717a670635a47c29e97d295044145"} Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.239796 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8011-account-create-hb7s5" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.249878 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0d75-account-create-kb2xh" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.254890 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c927-account-create-vwdhf" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.299354 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4w75\" (UniqueName: \"kubernetes.io/projected/1655641f-54ec-48c1-bbb6-21e801e6a54f-kube-api-access-v4w75\") pod \"1655641f-54ec-48c1-bbb6-21e801e6a54f\" (UID: \"1655641f-54ec-48c1-bbb6-21e801e6a54f\") " Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.299626 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jksbv\" (UniqueName: \"kubernetes.io/projected/7cf0be59-5a06-45c5-b913-d75c1e4096c0-kube-api-access-jksbv\") pod \"7cf0be59-5a06-45c5-b913-d75c1e4096c0\" (UID: \"7cf0be59-5a06-45c5-b913-d75c1e4096c0\") " Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.299818 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6z9j\" (UniqueName: \"kubernetes.io/projected/a7d44061-2a00-4a40-94c7-0c56e37d55c9-kube-api-access-l6z9j\") pod \"a7d44061-2a00-4a40-94c7-0c56e37d55c9\" (UID: \"a7d44061-2a00-4a40-94c7-0c56e37d55c9\") " Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.305444 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7d44061-2a00-4a40-94c7-0c56e37d55c9-kube-api-access-l6z9j" (OuterVolumeSpecName: "kube-api-access-l6z9j") pod "a7d44061-2a00-4a40-94c7-0c56e37d55c9" (UID: "a7d44061-2a00-4a40-94c7-0c56e37d55c9"). InnerVolumeSpecName "kube-api-access-l6z9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.305503 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1655641f-54ec-48c1-bbb6-21e801e6a54f-kube-api-access-v4w75" (OuterVolumeSpecName: "kube-api-access-v4w75") pod "1655641f-54ec-48c1-bbb6-21e801e6a54f" (UID: "1655641f-54ec-48c1-bbb6-21e801e6a54f"). InnerVolumeSpecName "kube-api-access-v4w75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.310442 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf0be59-5a06-45c5-b913-d75c1e4096c0-kube-api-access-jksbv" (OuterVolumeSpecName: "kube-api-access-jksbv") pod "7cf0be59-5a06-45c5-b913-d75c1e4096c0" (UID: "7cf0be59-5a06-45c5-b913-d75c1e4096c0"). InnerVolumeSpecName "kube-api-access-jksbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.402401 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6z9j\" (UniqueName: \"kubernetes.io/projected/a7d44061-2a00-4a40-94c7-0c56e37d55c9-kube-api-access-l6z9j\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.402447 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4w75\" (UniqueName: \"kubernetes.io/projected/1655641f-54ec-48c1-bbb6-21e801e6a54f-kube-api-access-v4w75\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.402459 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jksbv\" (UniqueName: \"kubernetes.io/projected/7cf0be59-5a06-45c5-b913-d75c1e4096c0-kube-api-access-jksbv\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.851232 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8011-account-create-hb7s5" event={"ID":"1655641f-54ec-48c1-bbb6-21e801e6a54f","Type":"ContainerDied","Data":"85e356354f9b0cdddf2dab4d7c008dc38900266d074fdf8ef5adaaefdb621723"} Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.851277 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85e356354f9b0cdddf2dab4d7c008dc38900266d074fdf8ef5adaaefdb621723" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.851254 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8011-account-create-hb7s5" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.852568 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0d75-account-create-kb2xh" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.852583 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0d75-account-create-kb2xh" event={"ID":"a7d44061-2a00-4a40-94c7-0c56e37d55c9","Type":"ContainerDied","Data":"88f36a385e300d49eff13c35709907f2489717a670635a47c29e97d295044145"} Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.852618 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88f36a385e300d49eff13c35709907f2489717a670635a47c29e97d295044145" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.854200 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c927-account-create-vwdhf" event={"ID":"7cf0be59-5a06-45c5-b913-d75c1e4096c0","Type":"ContainerDied","Data":"69f0a676e849637f6e898fe0d903c380d6115091913dd03c0eb8081c0bba151a"} Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.854224 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69f0a676e849637f6e898fe0d903c380d6115091913dd03c0eb8081c0bba151a" Oct 10 16:19:09 crc kubenswrapper[4788]: I1010 16:19:09.854272 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c927-account-create-vwdhf" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.448468 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w972c"] Oct 10 16:19:11 crc kubenswrapper[4788]: E1010 16:19:11.449172 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7d44061-2a00-4a40-94c7-0c56e37d55c9" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.449184 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7d44061-2a00-4a40-94c7-0c56e37d55c9" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: E1010 16:19:11.449219 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1655641f-54ec-48c1-bbb6-21e801e6a54f" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.449225 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1655641f-54ec-48c1-bbb6-21e801e6a54f" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: E1010 16:19:11.449233 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cf0be59-5a06-45c5-b913-d75c1e4096c0" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.449240 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cf0be59-5a06-45c5-b913-d75c1e4096c0" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.449392 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7d44061-2a00-4a40-94c7-0c56e37d55c9" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.449410 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1655641f-54ec-48c1-bbb6-21e801e6a54f" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.449420 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cf0be59-5a06-45c5-b913-d75c1e4096c0" containerName="mariadb-account-create" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.450053 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.454657 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.454846 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.454876 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-knttl" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.458790 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w972c"] Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.539682 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.539827 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-config-data\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.539903 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg9lh\" (UniqueName: \"kubernetes.io/projected/06b3f17f-fbea-4f87-b0eb-4fc601596d51-kube-api-access-pg9lh\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.539946 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-scripts\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.641643 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.641754 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-config-data\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.641787 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg9lh\" (UniqueName: \"kubernetes.io/projected/06b3f17f-fbea-4f87-b0eb-4fc601596d51-kube-api-access-pg9lh\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.641817 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-scripts\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.648801 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.651579 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-scripts\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.652418 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-config-data\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.657481 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg9lh\" (UniqueName: \"kubernetes.io/projected/06b3f17f-fbea-4f87-b0eb-4fc601596d51-kube-api-access-pg9lh\") pod \"nova-cell0-conductor-db-sync-w972c\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:11 crc kubenswrapper[4788]: I1010 16:19:11.776900 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:12 crc kubenswrapper[4788]: I1010 16:19:12.211215 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w972c"] Oct 10 16:19:12 crc kubenswrapper[4788]: I1010 16:19:12.882709 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-w972c" event={"ID":"06b3f17f-fbea-4f87-b0eb-4fc601596d51","Type":"ContainerStarted","Data":"4d1039c1755f06c7ab69ba240ecbd4b8f78e464881ef87611b32129e4152d420"} Oct 10 16:19:12 crc kubenswrapper[4788]: I1010 16:19:12.882754 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-w972c" event={"ID":"06b3f17f-fbea-4f87-b0eb-4fc601596d51","Type":"ContainerStarted","Data":"a29eed25ffdacbb85b31f2f9fed67fc6bc4ea357f6697c690d186b36cbd64fa6"} Oct 10 16:19:12 crc kubenswrapper[4788]: I1010 16:19:12.900040 4788 scope.go:117] "RemoveContainer" containerID="1c6717185a7b9e9ecdd28f9360f13a808cd80aa46ad52be555766c9af8df9ac1" Oct 10 16:19:12 crc kubenswrapper[4788]: I1010 16:19:12.903431 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-w972c" podStartSLOduration=1.903413221 podStartE2EDuration="1.903413221s" podCreationTimestamp="2025-10-10 16:19:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:12.896446332 +0000 UTC m=+5655.346161880" watchObservedRunningTime="2025-10-10 16:19:12.903413221 +0000 UTC m=+5655.353128769" Oct 10 16:19:17 crc kubenswrapper[4788]: I1010 16:19:17.954438 4788 generic.go:334] "Generic (PLEG): container finished" podID="06b3f17f-fbea-4f87-b0eb-4fc601596d51" containerID="4d1039c1755f06c7ab69ba240ecbd4b8f78e464881ef87611b32129e4152d420" exitCode=0 Oct 10 16:19:17 crc kubenswrapper[4788]: I1010 16:19:17.954499 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-w972c" event={"ID":"06b3f17f-fbea-4f87-b0eb-4fc601596d51","Type":"ContainerDied","Data":"4d1039c1755f06c7ab69ba240ecbd4b8f78e464881ef87611b32129e4152d420"} Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.235076 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:19:19 crc kubenswrapper[4788]: E1010 16:19:19.236113 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.372954 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.426199 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-scripts\") pod \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.426283 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg9lh\" (UniqueName: \"kubernetes.io/projected/06b3f17f-fbea-4f87-b0eb-4fc601596d51-kube-api-access-pg9lh\") pod \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.426427 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-combined-ca-bundle\") pod \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.426540 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-config-data\") pod \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\" (UID: \"06b3f17f-fbea-4f87-b0eb-4fc601596d51\") " Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.447753 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06b3f17f-fbea-4f87-b0eb-4fc601596d51-kube-api-access-pg9lh" (OuterVolumeSpecName: "kube-api-access-pg9lh") pod "06b3f17f-fbea-4f87-b0eb-4fc601596d51" (UID: "06b3f17f-fbea-4f87-b0eb-4fc601596d51"). InnerVolumeSpecName "kube-api-access-pg9lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.448128 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-scripts" (OuterVolumeSpecName: "scripts") pod "06b3f17f-fbea-4f87-b0eb-4fc601596d51" (UID: "06b3f17f-fbea-4f87-b0eb-4fc601596d51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.464850 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06b3f17f-fbea-4f87-b0eb-4fc601596d51" (UID: "06b3f17f-fbea-4f87-b0eb-4fc601596d51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.469773 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-config-data" (OuterVolumeSpecName: "config-data") pod "06b3f17f-fbea-4f87-b0eb-4fc601596d51" (UID: "06b3f17f-fbea-4f87-b0eb-4fc601596d51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.529410 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.529456 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg9lh\" (UniqueName: \"kubernetes.io/projected/06b3f17f-fbea-4f87-b0eb-4fc601596d51-kube-api-access-pg9lh\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.529474 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.529487 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06b3f17f-fbea-4f87-b0eb-4fc601596d51-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.984960 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-w972c" event={"ID":"06b3f17f-fbea-4f87-b0eb-4fc601596d51","Type":"ContainerDied","Data":"a29eed25ffdacbb85b31f2f9fed67fc6bc4ea357f6697c690d186b36cbd64fa6"} Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.985009 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a29eed25ffdacbb85b31f2f9fed67fc6bc4ea357f6697c690d186b36cbd64fa6" Oct 10 16:19:19 crc kubenswrapper[4788]: I1010 16:19:19.985066 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-w972c" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.065681 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:19:20 crc kubenswrapper[4788]: E1010 16:19:20.066221 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b3f17f-fbea-4f87-b0eb-4fc601596d51" containerName="nova-cell0-conductor-db-sync" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.066241 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b3f17f-fbea-4f87-b0eb-4fc601596d51" containerName="nova-cell0-conductor-db-sync" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.066437 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="06b3f17f-fbea-4f87-b0eb-4fc601596d51" containerName="nova-cell0-conductor-db-sync" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.067196 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.072201 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-knttl" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.072841 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.080679 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.142996 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.143089 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf9zt\" (UniqueName: \"kubernetes.io/projected/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-kube-api-access-pf9zt\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.143250 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.245280 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.245413 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.245638 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf9zt\" (UniqueName: \"kubernetes.io/projected/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-kube-api-access-pf9zt\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.251963 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.259357 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.262970 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf9zt\" (UniqueName: \"kubernetes.io/projected/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-kube-api-access-pf9zt\") pod \"nova-cell0-conductor-0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.419806 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.669214 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.997520 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0","Type":"ContainerStarted","Data":"3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04"} Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.997964 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0","Type":"ContainerStarted","Data":"e368c3ede74b655d93a69dfd0bca29af6f185d0a2d4872f06c5589872646fa78"} Oct 10 16:19:20 crc kubenswrapper[4788]: I1010 16:19:20.997987 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:21 crc kubenswrapper[4788]: I1010 16:19:21.020053 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.020036106 podStartE2EDuration="1.020036106s" podCreationTimestamp="2025-10-10 16:19:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:21.016093734 +0000 UTC m=+5663.465809282" watchObservedRunningTime="2025-10-10 16:19:21.020036106 +0000 UTC m=+5663.469751654" Oct 10 16:19:25 crc kubenswrapper[4788]: I1010 16:19:25.476844 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.123249 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-q4s2n"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.125026 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.128879 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.145239 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.180701 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.180815 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-scripts\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.180867 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-config-data\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.180920 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q5fr\" (UniqueName: \"kubernetes.io/projected/d29917ed-6776-4a18-8b81-e1034a2f68f4-kube-api-access-6q5fr\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.185090 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-q4s2n"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.262289 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.267387 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.273330 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.282333 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.282414 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-scripts\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.282455 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-config-data\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.282495 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q5fr\" (UniqueName: \"kubernetes.io/projected/d29917ed-6776-4a18-8b81-e1034a2f68f4-kube-api-access-6q5fr\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.294878 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.297897 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-config-data\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.302475 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-scripts\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.308288 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.312046 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q5fr\" (UniqueName: \"kubernetes.io/projected/d29917ed-6776-4a18-8b81-e1034a2f68f4-kube-api-access-6q5fr\") pod \"nova-cell0-cell-mapping-q4s2n\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.338258 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.340761 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.351599 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.353621 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.373862 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.384044 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.384099 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw6zl\" (UniqueName: \"kubernetes.io/projected/953be95c-ef20-4799-90c0-e5cfd13b11e8-kube-api-access-xw6zl\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.384167 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmvgq\" (UniqueName: \"kubernetes.io/projected/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-kube-api-access-hmvgq\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.384188 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.384210 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/953be95c-ef20-4799-90c0-e5cfd13b11e8-logs\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.384237 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.384274 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-config-data\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.407494 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.413770 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.431712 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.462647 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.463839 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.468494 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.473926 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489348 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489552 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489617 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw6zl\" (UniqueName: \"kubernetes.io/projected/953be95c-ef20-4799-90c0-e5cfd13b11e8-kube-api-access-xw6zl\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489651 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489672 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-logs\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489709 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmvgq\" (UniqueName: \"kubernetes.io/projected/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-kube-api-access-hmvgq\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489729 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489746 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/953be95c-ef20-4799-90c0-e5cfd13b11e8-logs\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489773 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489799 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-config-data\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489827 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-config-data\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.489871 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6z7s\" (UniqueName: \"kubernetes.io/projected/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-kube-api-access-v6z7s\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.495315 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.497294 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/953be95c-ef20-4799-90c0-e5cfd13b11e8-logs\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.508120 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.508361 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-config-data\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.508851 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.513974 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fd4f44ddf-4qddw"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.515543 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.518427 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmvgq\" (UniqueName: \"kubernetes.io/projected/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-kube-api-access-hmvgq\") pod \"nova-cell1-novncproxy-0\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.522011 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw6zl\" (UniqueName: \"kubernetes.io/projected/953be95c-ef20-4799-90c0-e5cfd13b11e8-kube-api-access-xw6zl\") pod \"nova-metadata-0\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.530622 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fd4f44ddf-4qddw"] Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.592218 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6z7s\" (UniqueName: \"kubernetes.io/projected/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-kube-api-access-v6z7s\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.592550 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-config-data\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.598633 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-dns-svc\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.598866 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.598921 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-sb\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.598952 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bmph\" (UniqueName: \"kubernetes.io/projected/ea15b81b-bb54-40bf-a470-70645b0e7759-kube-api-access-6bmph\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.598974 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-logs\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.599018 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-nb\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.599089 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-config\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.599189 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-config-data\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.599244 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.599280 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxfgw\" (UniqueName: \"kubernetes.io/projected/16151fa2-9a89-4a5e-a1c7-4853427e8851-kube-api-access-wxfgw\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.600900 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-logs\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.604772 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.606701 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-config-data\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.616091 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6z7s\" (UniqueName: \"kubernetes.io/projected/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-kube-api-access-v6z7s\") pod \"nova-api-0\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.694578 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.701508 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-sb\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.701565 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bmph\" (UniqueName: \"kubernetes.io/projected/ea15b81b-bb54-40bf-a470-70645b0e7759-kube-api-access-6bmph\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.701599 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-nb\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.701637 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-config\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.701698 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.701720 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxfgw\" (UniqueName: \"kubernetes.io/projected/16151fa2-9a89-4a5e-a1c7-4853427e8851-kube-api-access-wxfgw\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.701756 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-config-data\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.701776 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-dns-svc\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.704501 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-dns-svc\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.704598 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-nb\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.706318 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-sb\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.706457 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-config\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.707412 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-config-data\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.707930 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.715448 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.720160 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxfgw\" (UniqueName: \"kubernetes.io/projected/16151fa2-9a89-4a5e-a1c7-4853427e8851-kube-api-access-wxfgw\") pod \"nova-scheduler-0\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.723404 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bmph\" (UniqueName: \"kubernetes.io/projected/ea15b81b-bb54-40bf-a470-70645b0e7759-kube-api-access-6bmph\") pod \"dnsmasq-dns-fd4f44ddf-4qddw\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.748838 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.926293 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.935894 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:26 crc kubenswrapper[4788]: I1010 16:19:26.992073 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-q4s2n"] Oct 10 16:19:27 crc kubenswrapper[4788]: W1010 16:19:27.028420 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd29917ed_6776_4a18_8b81_e1034a2f68f4.slice/crio-8e1ae329ead421d5e1b35bc0349689e662c60147202e90f4515b0eb2aebe5d90 WatchSource:0}: Error finding container 8e1ae329ead421d5e1b35bc0349689e662c60147202e90f4515b0eb2aebe5d90: Status 404 returned error can't find the container with id 8e1ae329ead421d5e1b35bc0349689e662c60147202e90f4515b0eb2aebe5d90 Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.095572 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q4s2n" event={"ID":"d29917ed-6776-4a18-8b81-e1034a2f68f4","Type":"ContainerStarted","Data":"8e1ae329ead421d5e1b35bc0349689e662c60147202e90f4515b0eb2aebe5d90"} Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.287462 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:27 crc kubenswrapper[4788]: W1010 16:19:27.293288 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod953be95c_ef20_4799_90c0_e5cfd13b11e8.slice/crio-b859dc25e6ae7b8467db963dd40691b3472e69996a19136f86336eb6ecd971ff WatchSource:0}: Error finding container b859dc25e6ae7b8467db963dd40691b3472e69996a19136f86336eb6ecd971ff: Status 404 returned error can't find the container with id b859dc25e6ae7b8467db963dd40691b3472e69996a19136f86336eb6ecd971ff Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.354215 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.445668 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.543298 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.556721 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fd4f44ddf-4qddw"] Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.709568 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2pdgt"] Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.711344 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.716088 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.716514 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.730081 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2pdgt"] Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.837356 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mq7p\" (UniqueName: \"kubernetes.io/projected/a1cff6f4-b4e1-420d-9723-25d83b9ee079-kube-api-access-6mq7p\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.837683 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-scripts\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.837918 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-config-data\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.838211 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.939466 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mq7p\" (UniqueName: \"kubernetes.io/projected/a1cff6f4-b4e1-420d-9723-25d83b9ee079-kube-api-access-6mq7p\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.939563 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-scripts\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.939630 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-config-data\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.939700 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.944551 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-scripts\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.944816 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.944863 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-config-data\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:27 crc kubenswrapper[4788]: I1010 16:19:27.959860 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mq7p\" (UniqueName: \"kubernetes.io/projected/a1cff6f4-b4e1-420d-9723-25d83b9ee079-kube-api-access-6mq7p\") pod \"nova-cell1-conductor-db-sync-2pdgt\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.056242 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.106058 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q4s2n" event={"ID":"d29917ed-6776-4a18-8b81-e1034a2f68f4","Type":"ContainerStarted","Data":"aab30d0014507e0985e587fd8c408952d9eaf41ca860494081804986a1d0230e"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.112201 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"16151fa2-9a89-4a5e-a1c7-4853427e8851","Type":"ContainerStarted","Data":"9079b3ba03acb325eff3c901f40d1fb3168c67d2061b3cb665d12702d2494d96"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.112245 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"16151fa2-9a89-4a5e-a1c7-4853427e8851","Type":"ContainerStarted","Data":"c335ec8c4bc35793fc142444eccdf2d65dec60f7500f02b73f28c70900c8b8a7"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.114866 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e","Type":"ContainerStarted","Data":"0ae6b8020e8b0a2f6507160ffdffb85d8c06bde17f8669da74d72ea84027de67"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.114914 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e","Type":"ContainerStarted","Data":"ea15d58e56b69d6a65d4ea2fd670d21633ac37800ac100c28176e6a349971c1d"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.116558 4788 generic.go:334] "Generic (PLEG): container finished" podID="ea15b81b-bb54-40bf-a470-70645b0e7759" containerID="a713fb395ae24a60f376ea4be0235b6027261ee459409d2bb69aa57af3d2ede5" exitCode=0 Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.116627 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" event={"ID":"ea15b81b-bb54-40bf-a470-70645b0e7759","Type":"ContainerDied","Data":"a713fb395ae24a60f376ea4be0235b6027261ee459409d2bb69aa57af3d2ede5"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.116656 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" event={"ID":"ea15b81b-bb54-40bf-a470-70645b0e7759","Type":"ContainerStarted","Data":"4af723edf17f9055636caf0225801764f3c3a3a9bb7a206d235a17c11372b21e"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.119509 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72c62b28-a8ed-4e80-967f-7a51bd5f4d28","Type":"ContainerStarted","Data":"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.119550 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72c62b28-a8ed-4e80-967f-7a51bd5f4d28","Type":"ContainerStarted","Data":"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.119563 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72c62b28-a8ed-4e80-967f-7a51bd5f4d28","Type":"ContainerStarted","Data":"dc4f6d39e2c5038e2da3eb7d7139456cc141fba0544e9497ccacd828b7371632"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.134924 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"953be95c-ef20-4799-90c0-e5cfd13b11e8","Type":"ContainerStarted","Data":"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.134985 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"953be95c-ef20-4799-90c0-e5cfd13b11e8","Type":"ContainerStarted","Data":"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.134998 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"953be95c-ef20-4799-90c0-e5cfd13b11e8","Type":"ContainerStarted","Data":"b859dc25e6ae7b8467db963dd40691b3472e69996a19136f86336eb6ecd971ff"} Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.152796 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-q4s2n" podStartSLOduration=2.152771943 podStartE2EDuration="2.152771943s" podCreationTimestamp="2025-10-10 16:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:28.124549369 +0000 UTC m=+5670.574264917" watchObservedRunningTime="2025-10-10 16:19:28.152771943 +0000 UTC m=+5670.602487491" Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.163192 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.1631727 podStartE2EDuration="2.1631727s" podCreationTimestamp="2025-10-10 16:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:28.143745172 +0000 UTC m=+5670.593460710" watchObservedRunningTime="2025-10-10 16:19:28.1631727 +0000 UTC m=+5670.612888248" Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.227060 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.227042781 podStartE2EDuration="2.227042781s" podCreationTimestamp="2025-10-10 16:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:28.207315795 +0000 UTC m=+5670.657031343" watchObservedRunningTime="2025-10-10 16:19:28.227042781 +0000 UTC m=+5670.676758329" Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.247220 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.247201379 podStartE2EDuration="2.247201379s" podCreationTimestamp="2025-10-10 16:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:28.243251958 +0000 UTC m=+5670.692967516" watchObservedRunningTime="2025-10-10 16:19:28.247201379 +0000 UTC m=+5670.696916927" Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.343509 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.343223386 podStartE2EDuration="2.343223386s" podCreationTimestamp="2025-10-10 16:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:28.316561561 +0000 UTC m=+5670.766277109" watchObservedRunningTime="2025-10-10 16:19:28.343223386 +0000 UTC m=+5670.792938944" Oct 10 16:19:28 crc kubenswrapper[4788]: I1010 16:19:28.541779 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2pdgt"] Oct 10 16:19:29 crc kubenswrapper[4788]: I1010 16:19:29.162676 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" event={"ID":"ea15b81b-bb54-40bf-a470-70645b0e7759","Type":"ContainerStarted","Data":"fd9d27226e6c45b71cbe8ed9cd691fa7779d8477be4919e335e7fafb8bb94218"} Oct 10 16:19:29 crc kubenswrapper[4788]: I1010 16:19:29.164213 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:29 crc kubenswrapper[4788]: I1010 16:19:29.166971 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" event={"ID":"a1cff6f4-b4e1-420d-9723-25d83b9ee079","Type":"ContainerStarted","Data":"7e3baf9d30c8a0b8a0496f374a184959c6cee6152f43000611a26b5dcd775b30"} Oct 10 16:19:29 crc kubenswrapper[4788]: I1010 16:19:29.166999 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" event={"ID":"a1cff6f4-b4e1-420d-9723-25d83b9ee079","Type":"ContainerStarted","Data":"9a31f6febe0d29ccb9091f825967f706752208ac398ed3eac9ecf15ce0d85947"} Oct 10 16:19:29 crc kubenswrapper[4788]: I1010 16:19:29.187830 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" podStartSLOduration=3.187814353 podStartE2EDuration="3.187814353s" podCreationTimestamp="2025-10-10 16:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:29.186606512 +0000 UTC m=+5671.636322060" watchObservedRunningTime="2025-10-10 16:19:29.187814353 +0000 UTC m=+5671.637529901" Oct 10 16:19:29 crc kubenswrapper[4788]: I1010 16:19:29.205514 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" podStartSLOduration=2.205500488 podStartE2EDuration="2.205500488s" podCreationTimestamp="2025-10-10 16:19:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:29.201249799 +0000 UTC m=+5671.650965347" watchObservedRunningTime="2025-10-10 16:19:29.205500488 +0000 UTC m=+5671.655216036" Oct 10 16:19:31 crc kubenswrapper[4788]: I1010 16:19:31.695324 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:31 crc kubenswrapper[4788]: I1010 16:19:31.718549 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 16:19:31 crc kubenswrapper[4788]: I1010 16:19:31.718632 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 16:19:31 crc kubenswrapper[4788]: I1010 16:19:31.927821 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 16:19:32 crc kubenswrapper[4788]: I1010 16:19:32.208923 4788 generic.go:334] "Generic (PLEG): container finished" podID="a1cff6f4-b4e1-420d-9723-25d83b9ee079" containerID="7e3baf9d30c8a0b8a0496f374a184959c6cee6152f43000611a26b5dcd775b30" exitCode=0 Oct 10 16:19:32 crc kubenswrapper[4788]: I1010 16:19:32.209213 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" event={"ID":"a1cff6f4-b4e1-420d-9723-25d83b9ee079","Type":"ContainerDied","Data":"7e3baf9d30c8a0b8a0496f374a184959c6cee6152f43000611a26b5dcd775b30"} Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.225802 4788 generic.go:334] "Generic (PLEG): container finished" podID="d29917ed-6776-4a18-8b81-e1034a2f68f4" containerID="aab30d0014507e0985e587fd8c408952d9eaf41ca860494081804986a1d0230e" exitCode=0 Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.225899 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q4s2n" event={"ID":"d29917ed-6776-4a18-8b81-e1034a2f68f4","Type":"ContainerDied","Data":"aab30d0014507e0985e587fd8c408952d9eaf41ca860494081804986a1d0230e"} Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.234991 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:19:33 crc kubenswrapper[4788]: E1010 16:19:33.235542 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.604850 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.764975 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-scripts\") pod \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.765329 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-combined-ca-bundle\") pod \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.765428 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-config-data\") pod \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.765518 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mq7p\" (UniqueName: \"kubernetes.io/projected/a1cff6f4-b4e1-420d-9723-25d83b9ee079-kube-api-access-6mq7p\") pod \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\" (UID: \"a1cff6f4-b4e1-420d-9723-25d83b9ee079\") " Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.772014 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1cff6f4-b4e1-420d-9723-25d83b9ee079-kube-api-access-6mq7p" (OuterVolumeSpecName: "kube-api-access-6mq7p") pod "a1cff6f4-b4e1-420d-9723-25d83b9ee079" (UID: "a1cff6f4-b4e1-420d-9723-25d83b9ee079"). InnerVolumeSpecName "kube-api-access-6mq7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.781476 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-scripts" (OuterVolumeSpecName: "scripts") pod "a1cff6f4-b4e1-420d-9723-25d83b9ee079" (UID: "a1cff6f4-b4e1-420d-9723-25d83b9ee079"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.794648 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-config-data" (OuterVolumeSpecName: "config-data") pod "a1cff6f4-b4e1-420d-9723-25d83b9ee079" (UID: "a1cff6f4-b4e1-420d-9723-25d83b9ee079"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.795601 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1cff6f4-b4e1-420d-9723-25d83b9ee079" (UID: "a1cff6f4-b4e1-420d-9723-25d83b9ee079"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.869075 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.869115 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.869130 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1cff6f4-b4e1-420d-9723-25d83b9ee079-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:33 crc kubenswrapper[4788]: I1010 16:19:33.869165 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mq7p\" (UniqueName: \"kubernetes.io/projected/a1cff6f4-b4e1-420d-9723-25d83b9ee079-kube-api-access-6mq7p\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.239681 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.272379 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2pdgt" event={"ID":"a1cff6f4-b4e1-420d-9723-25d83b9ee079","Type":"ContainerDied","Data":"9a31f6febe0d29ccb9091f825967f706752208ac398ed3eac9ecf15ce0d85947"} Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.272568 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a31f6febe0d29ccb9091f825967f706752208ac398ed3eac9ecf15ce0d85947" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.315953 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:19:34 crc kubenswrapper[4788]: E1010 16:19:34.318564 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1cff6f4-b4e1-420d-9723-25d83b9ee079" containerName="nova-cell1-conductor-db-sync" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.318593 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1cff6f4-b4e1-420d-9723-25d83b9ee079" containerName="nova-cell1-conductor-db-sync" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.318806 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1cff6f4-b4e1-420d-9723-25d83b9ee079" containerName="nova-cell1-conductor-db-sync" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.319833 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.324945 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.332783 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.378321 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.378635 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.378688 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl7z8\" (UniqueName: \"kubernetes.io/projected/bfe95d07-0b8f-49cf-9bb4-6817e193770e-kube-api-access-gl7z8\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.480105 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.480190 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl7z8\" (UniqueName: \"kubernetes.io/projected/bfe95d07-0b8f-49cf-9bb4-6817e193770e-kube-api-access-gl7z8\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.480220 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.486681 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.488523 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.501654 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl7z8\" (UniqueName: \"kubernetes.io/projected/bfe95d07-0b8f-49cf-9bb4-6817e193770e-kube-api-access-gl7z8\") pod \"nova-cell1-conductor-0\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.634760 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.642266 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.796030 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-combined-ca-bundle\") pod \"d29917ed-6776-4a18-8b81-e1034a2f68f4\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.796259 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q5fr\" (UniqueName: \"kubernetes.io/projected/d29917ed-6776-4a18-8b81-e1034a2f68f4-kube-api-access-6q5fr\") pod \"d29917ed-6776-4a18-8b81-e1034a2f68f4\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.797045 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-config-data\") pod \"d29917ed-6776-4a18-8b81-e1034a2f68f4\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.797185 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-scripts\") pod \"d29917ed-6776-4a18-8b81-e1034a2f68f4\" (UID: \"d29917ed-6776-4a18-8b81-e1034a2f68f4\") " Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.802808 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-scripts" (OuterVolumeSpecName: "scripts") pod "d29917ed-6776-4a18-8b81-e1034a2f68f4" (UID: "d29917ed-6776-4a18-8b81-e1034a2f68f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.803234 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d29917ed-6776-4a18-8b81-e1034a2f68f4-kube-api-access-6q5fr" (OuterVolumeSpecName: "kube-api-access-6q5fr") pod "d29917ed-6776-4a18-8b81-e1034a2f68f4" (UID: "d29917ed-6776-4a18-8b81-e1034a2f68f4"). InnerVolumeSpecName "kube-api-access-6q5fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.830681 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d29917ed-6776-4a18-8b81-e1034a2f68f4" (UID: "d29917ed-6776-4a18-8b81-e1034a2f68f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.836227 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-config-data" (OuterVolumeSpecName: "config-data") pod "d29917ed-6776-4a18-8b81-e1034a2f68f4" (UID: "d29917ed-6776-4a18-8b81-e1034a2f68f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.900175 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q5fr\" (UniqueName: \"kubernetes.io/projected/d29917ed-6776-4a18-8b81-e1034a2f68f4-kube-api-access-6q5fr\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.900205 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.900217 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:34 crc kubenswrapper[4788]: I1010 16:19:34.900241 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29917ed-6776-4a18-8b81-e1034a2f68f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.138671 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.251421 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bfe95d07-0b8f-49cf-9bb4-6817e193770e","Type":"ContainerStarted","Data":"60b9dc4c95e75c0e30917f0275b33cbecc0bf12a9c3ccb49377baa380f4c4179"} Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.255297 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q4s2n" event={"ID":"d29917ed-6776-4a18-8b81-e1034a2f68f4","Type":"ContainerDied","Data":"8e1ae329ead421d5e1b35bc0349689e662c60147202e90f4515b0eb2aebe5d90"} Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.255341 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e1ae329ead421d5e1b35bc0349689e662c60147202e90f4515b0eb2aebe5d90" Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.255412 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q4s2n" Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.461775 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.462121 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerName="nova-api-log" containerID="cri-o://691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5" gracePeriod=30 Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.462247 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerName="nova-api-api" containerID="cri-o://cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5" gracePeriod=30 Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.475036 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.475452 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="16151fa2-9a89-4a5e-a1c7-4853427e8851" containerName="nova-scheduler-scheduler" containerID="cri-o://9079b3ba03acb325eff3c901f40d1fb3168c67d2061b3cb665d12702d2494d96" gracePeriod=30 Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.502005 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.502476 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerName="nova-metadata-log" containerID="cri-o://00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e" gracePeriod=30 Oct 10 16:19:35 crc kubenswrapper[4788]: I1010 16:19:35.502720 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerName="nova-metadata-metadata" containerID="cri-o://ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0" gracePeriod=30 Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.117522 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.130723 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.232830 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-config-data\") pod \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.232996 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6z7s\" (UniqueName: \"kubernetes.io/projected/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-kube-api-access-v6z7s\") pod \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.233210 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-logs\") pod \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.233380 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw6zl\" (UniqueName: \"kubernetes.io/projected/953be95c-ef20-4799-90c0-e5cfd13b11e8-kube-api-access-xw6zl\") pod \"953be95c-ef20-4799-90c0-e5cfd13b11e8\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.233487 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-combined-ca-bundle\") pod \"953be95c-ef20-4799-90c0-e5cfd13b11e8\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.233925 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-combined-ca-bundle\") pod \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\" (UID: \"72c62b28-a8ed-4e80-967f-7a51bd5f4d28\") " Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.234258 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-logs" (OuterVolumeSpecName: "logs") pod "72c62b28-a8ed-4e80-967f-7a51bd5f4d28" (UID: "72c62b28-a8ed-4e80-967f-7a51bd5f4d28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.234770 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.238933 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-kube-api-access-v6z7s" (OuterVolumeSpecName: "kube-api-access-v6z7s") pod "72c62b28-a8ed-4e80-967f-7a51bd5f4d28" (UID: "72c62b28-a8ed-4e80-967f-7a51bd5f4d28"). InnerVolumeSpecName "kube-api-access-v6z7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.240541 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/953be95c-ef20-4799-90c0-e5cfd13b11e8-kube-api-access-xw6zl" (OuterVolumeSpecName: "kube-api-access-xw6zl") pod "953be95c-ef20-4799-90c0-e5cfd13b11e8" (UID: "953be95c-ef20-4799-90c0-e5cfd13b11e8"). InnerVolumeSpecName "kube-api-access-xw6zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.259804 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-config-data" (OuterVolumeSpecName: "config-data") pod "72c62b28-a8ed-4e80-967f-7a51bd5f4d28" (UID: "72c62b28-a8ed-4e80-967f-7a51bd5f4d28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.268886 4788 generic.go:334] "Generic (PLEG): container finished" podID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerID="cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5" exitCode=0 Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.269110 4788 generic.go:334] "Generic (PLEG): container finished" podID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerID="691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5" exitCode=143 Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.268971 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72c62b28-a8ed-4e80-967f-7a51bd5f4d28","Type":"ContainerDied","Data":"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5"} Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.269350 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72c62b28-a8ed-4e80-967f-7a51bd5f4d28","Type":"ContainerDied","Data":"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5"} Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.269409 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"72c62b28-a8ed-4e80-967f-7a51bd5f4d28","Type":"ContainerDied","Data":"dc4f6d39e2c5038e2da3eb7d7139456cc141fba0544e9497ccacd828b7371632"} Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.268953 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.269444 4788 scope.go:117] "RemoveContainer" containerID="cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.272037 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bfe95d07-0b8f-49cf-9bb4-6817e193770e","Type":"ContainerStarted","Data":"44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251"} Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.272733 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.274921 4788 generic.go:334] "Generic (PLEG): container finished" podID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerID="ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0" exitCode=0 Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.274949 4788 generic.go:334] "Generic (PLEG): container finished" podID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerID="00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e" exitCode=143 Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.274972 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"953be95c-ef20-4799-90c0-e5cfd13b11e8","Type":"ContainerDied","Data":"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0"} Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.274995 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"953be95c-ef20-4799-90c0-e5cfd13b11e8","Type":"ContainerDied","Data":"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e"} Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.275006 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"953be95c-ef20-4799-90c0-e5cfd13b11e8","Type":"ContainerDied","Data":"b859dc25e6ae7b8467db963dd40691b3472e69996a19136f86336eb6ecd971ff"} Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.275061 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.276103 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72c62b28-a8ed-4e80-967f-7a51bd5f4d28" (UID: "72c62b28-a8ed-4e80-967f-7a51bd5f4d28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.280005 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "953be95c-ef20-4799-90c0-e5cfd13b11e8" (UID: "953be95c-ef20-4799-90c0-e5cfd13b11e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.297416 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.297397577 podStartE2EDuration="2.297397577s" podCreationTimestamp="2025-10-10 16:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:36.294070161 +0000 UTC m=+5678.743785719" watchObservedRunningTime="2025-10-10 16:19:36.297397577 +0000 UTC m=+5678.747113115" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.307376 4788 scope.go:117] "RemoveContainer" containerID="691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.323936 4788 scope.go:117] "RemoveContainer" containerID="cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5" Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.324343 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5\": container with ID starting with cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5 not found: ID does not exist" containerID="cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.324433 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5"} err="failed to get container status \"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5\": rpc error: code = NotFound desc = could not find container \"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5\": container with ID starting with cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5 not found: ID does not exist" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.324504 4788 scope.go:117] "RemoveContainer" containerID="691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5" Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.324913 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5\": container with ID starting with 691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5 not found: ID does not exist" containerID="691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.324958 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5"} err="failed to get container status \"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5\": rpc error: code = NotFound desc = could not find container \"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5\": container with ID starting with 691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5 not found: ID does not exist" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.324986 4788 scope.go:117] "RemoveContainer" containerID="cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.325318 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5"} err="failed to get container status \"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5\": rpc error: code = NotFound desc = could not find container \"cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5\": container with ID starting with cb342d6cc760977001223eca7582b2221b6cad36726271a377454948b62b01f5 not found: ID does not exist" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.325330 4788 scope.go:117] "RemoveContainer" containerID="691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.325511 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5"} err="failed to get container status \"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5\": rpc error: code = NotFound desc = could not find container \"691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5\": container with ID starting with 691daa431e2652308a7446582d5e083af17e696b4557482bd195f2a8878a97e5 not found: ID does not exist" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.325528 4788 scope.go:117] "RemoveContainer" containerID="ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.335295 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-config-data\") pod \"953be95c-ef20-4799-90c0-e5cfd13b11e8\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.335515 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/953be95c-ef20-4799-90c0-e5cfd13b11e8-logs\") pod \"953be95c-ef20-4799-90c0-e5cfd13b11e8\" (UID: \"953be95c-ef20-4799-90c0-e5cfd13b11e8\") " Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.335950 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.336396 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/953be95c-ef20-4799-90c0-e5cfd13b11e8-logs" (OuterVolumeSpecName: "logs") pod "953be95c-ef20-4799-90c0-e5cfd13b11e8" (UID: "953be95c-ef20-4799-90c0-e5cfd13b11e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.336525 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6z7s\" (UniqueName: \"kubernetes.io/projected/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-kube-api-access-v6z7s\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.338355 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw6zl\" (UniqueName: \"kubernetes.io/projected/953be95c-ef20-4799-90c0-e5cfd13b11e8-kube-api-access-xw6zl\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.338383 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.338397 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72c62b28-a8ed-4e80-967f-7a51bd5f4d28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.355405 4788 scope.go:117] "RemoveContainer" containerID="00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.370612 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-config-data" (OuterVolumeSpecName: "config-data") pod "953be95c-ef20-4799-90c0-e5cfd13b11e8" (UID: "953be95c-ef20-4799-90c0-e5cfd13b11e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.391119 4788 scope.go:117] "RemoveContainer" containerID="ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0" Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.391579 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0\": container with ID starting with ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0 not found: ID does not exist" containerID="ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.391608 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0"} err="failed to get container status \"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0\": rpc error: code = NotFound desc = could not find container \"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0\": container with ID starting with ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0 not found: ID does not exist" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.391633 4788 scope.go:117] "RemoveContainer" containerID="00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e" Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.391968 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e\": container with ID starting with 00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e not found: ID does not exist" containerID="00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.391989 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e"} err="failed to get container status \"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e\": rpc error: code = NotFound desc = could not find container \"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e\": container with ID starting with 00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e not found: ID does not exist" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.392003 4788 scope.go:117] "RemoveContainer" containerID="ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.392329 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0"} err="failed to get container status \"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0\": rpc error: code = NotFound desc = could not find container \"ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0\": container with ID starting with ef420adaba23e5fc5829a45a9c1bed97bc278d6062d93b1c79596da5a10b9dc0 not found: ID does not exist" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.392415 4788 scope.go:117] "RemoveContainer" containerID="00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.392762 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e"} err="failed to get container status \"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e\": rpc error: code = NotFound desc = could not find container \"00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e\": container with ID starting with 00d7d37b2ad35b3804072367afbb23c5a5054c990d3986efd2be08ce60f26a0e not found: ID does not exist" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.439805 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/953be95c-ef20-4799-90c0-e5cfd13b11e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.440098 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/953be95c-ef20-4799-90c0-e5cfd13b11e8-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.674850 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.684944 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.698323 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.706971 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.717946 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.723878 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.736242 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.736941 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerName="nova-api-log" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.736962 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerName="nova-api-log" Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.736979 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29917ed-6776-4a18-8b81-e1034a2f68f4" containerName="nova-manage" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.736988 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29917ed-6776-4a18-8b81-e1034a2f68f4" containerName="nova-manage" Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.737033 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerName="nova-metadata-log" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.737043 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerName="nova-metadata-log" Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.737062 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerName="nova-metadata-metadata" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.737069 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerName="nova-metadata-metadata" Oct 10 16:19:36 crc kubenswrapper[4788]: E1010 16:19:36.737088 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerName="nova-api-api" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.737098 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerName="nova-api-api" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.737352 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerName="nova-api-log" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.737375 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" containerName="nova-api-api" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.737392 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerName="nova-metadata-log" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.737408 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" containerName="nova-metadata-metadata" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.737425 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d29917ed-6776-4a18-8b81-e1034a2f68f4" containerName="nova-manage" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.742917 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.746524 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.786054 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.791445 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.794893 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.825176 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.836064 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.848132 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a531e3a-9184-4aa4-9a68-5857a2a3df19-logs\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.848203 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.848316 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pgvk\" (UniqueName: \"kubernetes.io/projected/0a531e3a-9184-4aa4-9a68-5857a2a3df19-kube-api-access-2pgvk\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.848343 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-config-data\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.938309 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.951469 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/944a91e4-1502-4abc-a51d-7fe2b61cb635-logs\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.951552 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xp5p\" (UniqueName: \"kubernetes.io/projected/944a91e4-1502-4abc-a51d-7fe2b61cb635-kube-api-access-8xp5p\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.951631 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.951665 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a531e3a-9184-4aa4-9a68-5857a2a3df19-logs\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.951716 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.951780 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-config-data\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.951820 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pgvk\" (UniqueName: \"kubernetes.io/projected/0a531e3a-9184-4aa4-9a68-5857a2a3df19-kube-api-access-2pgvk\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.951857 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-config-data\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.952791 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a531e3a-9184-4aa4-9a68-5857a2a3df19-logs\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.959066 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.980981 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-config-data\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:36 crc kubenswrapper[4788]: I1010 16:19:36.991940 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pgvk\" (UniqueName: \"kubernetes.io/projected/0a531e3a-9184-4aa4-9a68-5857a2a3df19-kube-api-access-2pgvk\") pod \"nova-api-0\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " pod="openstack/nova-api-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.040046 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bb448fdbf-sbkd8"] Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.040290 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" podUID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" containerName="dnsmasq-dns" containerID="cri-o://426a1f0c09305f48ce78f7012b0c38962132fd4e6911d21378d09517cf78cf68" gracePeriod=10 Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.056542 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.057064 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-config-data\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.057466 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/944a91e4-1502-4abc-a51d-7fe2b61cb635-logs\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.057514 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xp5p\" (UniqueName: \"kubernetes.io/projected/944a91e4-1502-4abc-a51d-7fe2b61cb635-kube-api-access-8xp5p\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.058040 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/944a91e4-1502-4abc-a51d-7fe2b61cb635-logs\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.060737 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-config-data\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.061217 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.074367 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.083878 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xp5p\" (UniqueName: \"kubernetes.io/projected/944a91e4-1502-4abc-a51d-7fe2b61cb635-kube-api-access-8xp5p\") pod \"nova-metadata-0\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.111916 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.307609 4788 generic.go:334] "Generic (PLEG): container finished" podID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" containerID="426a1f0c09305f48ce78f7012b0c38962132fd4e6911d21378d09517cf78cf68" exitCode=0 Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.307710 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" event={"ID":"8ea194fc-2dde-490a-b9d5-6f17dfd3af19","Type":"ContainerDied","Data":"426a1f0c09305f48ce78f7012b0c38962132fd4e6911d21378d09517cf78cf68"} Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.318411 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.493121 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.661941 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.669874 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-sb\") pod \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.669943 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-nb\") pod \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.670048 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkmh2\" (UniqueName: \"kubernetes.io/projected/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-kube-api-access-bkmh2\") pod \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.670099 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-dns-svc\") pod \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.670182 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-config\") pod \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\" (UID: \"8ea194fc-2dde-490a-b9d5-6f17dfd3af19\") " Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.676540 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-kube-api-access-bkmh2" (OuterVolumeSpecName: "kube-api-access-bkmh2") pod "8ea194fc-2dde-490a-b9d5-6f17dfd3af19" (UID: "8ea194fc-2dde-490a-b9d5-6f17dfd3af19"). InnerVolumeSpecName "kube-api-access-bkmh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.740729 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8ea194fc-2dde-490a-b9d5-6f17dfd3af19" (UID: "8ea194fc-2dde-490a-b9d5-6f17dfd3af19"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.748583 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-config" (OuterVolumeSpecName: "config") pod "8ea194fc-2dde-490a-b9d5-6f17dfd3af19" (UID: "8ea194fc-2dde-490a-b9d5-6f17dfd3af19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.760603 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.766263 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ea194fc-2dde-490a-b9d5-6f17dfd3af19" (UID: "8ea194fc-2dde-490a-b9d5-6f17dfd3af19"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.772313 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.772690 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkmh2\" (UniqueName: \"kubernetes.io/projected/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-kube-api-access-bkmh2\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.772823 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.772908 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.782975 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8ea194fc-2dde-490a-b9d5-6f17dfd3af19" (UID: "8ea194fc-2dde-490a-b9d5-6f17dfd3af19"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:19:37 crc kubenswrapper[4788]: I1010 16:19:37.874610 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ea194fc-2dde-490a-b9d5-6f17dfd3af19-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.251901 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72c62b28-a8ed-4e80-967f-7a51bd5f4d28" path="/var/lib/kubelet/pods/72c62b28-a8ed-4e80-967f-7a51bd5f4d28/volumes" Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.252536 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="953be95c-ef20-4799-90c0-e5cfd13b11e8" path="/var/lib/kubelet/pods/953be95c-ef20-4799-90c0-e5cfd13b11e8/volumes" Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.321260 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" event={"ID":"8ea194fc-2dde-490a-b9d5-6f17dfd3af19","Type":"ContainerDied","Data":"4e12fbe4a231c17d546ffdb2edbc42eafa2fe698d4cd7e3b78129a99ffb76cdb"} Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.321322 4788 scope.go:117] "RemoveContainer" containerID="426a1f0c09305f48ce78f7012b0c38962132fd4e6911d21378d09517cf78cf68" Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.321471 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bb448fdbf-sbkd8" Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.324766 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"944a91e4-1502-4abc-a51d-7fe2b61cb635","Type":"ContainerStarted","Data":"823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a"} Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.324814 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"944a91e4-1502-4abc-a51d-7fe2b61cb635","Type":"ContainerStarted","Data":"b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f"} Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.324827 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"944a91e4-1502-4abc-a51d-7fe2b61cb635","Type":"ContainerStarted","Data":"c93b6d04be9d870eb78885e36f50060c8e525e2a46ac202de0319dca96d31073"} Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.338375 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a531e3a-9184-4aa4-9a68-5857a2a3df19","Type":"ContainerStarted","Data":"97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1"} Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.338436 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a531e3a-9184-4aa4-9a68-5857a2a3df19","Type":"ContainerStarted","Data":"ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1"} Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.338451 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a531e3a-9184-4aa4-9a68-5857a2a3df19","Type":"ContainerStarted","Data":"e4720ccbc105b07726d490fbad296b36f2cb29e91cbd8d2ee07af7ea7845033c"} Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.343215 4788 scope.go:117] "RemoveContainer" containerID="ace6045c4c7d1dc4d05567790fea276f0de9572dd348e214200222f24131998d" Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.385121 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.385089149 podStartE2EDuration="2.385089149s" podCreationTimestamp="2025-10-10 16:19:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:38.356234437 +0000 UTC m=+5680.805949975" watchObservedRunningTime="2025-10-10 16:19:38.385089149 +0000 UTC m=+5680.834804697" Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.387380 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bb448fdbf-sbkd8"] Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.399541 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bb448fdbf-sbkd8"] Oct 10 16:19:38 crc kubenswrapper[4788]: I1010 16:19:38.406265 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.406240852 podStartE2EDuration="2.406240852s" podCreationTimestamp="2025-10-10 16:19:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:38.389104572 +0000 UTC m=+5680.838820130" watchObservedRunningTime="2025-10-10 16:19:38.406240852 +0000 UTC m=+5680.855956400" Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.355815 4788 generic.go:334] "Generic (PLEG): container finished" podID="16151fa2-9a89-4a5e-a1c7-4853427e8851" containerID="9079b3ba03acb325eff3c901f40d1fb3168c67d2061b3cb665d12702d2494d96" exitCode=0 Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.355926 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"16151fa2-9a89-4a5e-a1c7-4853427e8851","Type":"ContainerDied","Data":"9079b3ba03acb325eff3c901f40d1fb3168c67d2061b3cb665d12702d2494d96"} Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.675527 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.822636 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxfgw\" (UniqueName: \"kubernetes.io/projected/16151fa2-9a89-4a5e-a1c7-4853427e8851-kube-api-access-wxfgw\") pod \"16151fa2-9a89-4a5e-a1c7-4853427e8851\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.823009 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-config-data\") pod \"16151fa2-9a89-4a5e-a1c7-4853427e8851\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.823208 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-combined-ca-bundle\") pod \"16151fa2-9a89-4a5e-a1c7-4853427e8851\" (UID: \"16151fa2-9a89-4a5e-a1c7-4853427e8851\") " Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.828902 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16151fa2-9a89-4a5e-a1c7-4853427e8851-kube-api-access-wxfgw" (OuterVolumeSpecName: "kube-api-access-wxfgw") pod "16151fa2-9a89-4a5e-a1c7-4853427e8851" (UID: "16151fa2-9a89-4a5e-a1c7-4853427e8851"). InnerVolumeSpecName "kube-api-access-wxfgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.853305 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-config-data" (OuterVolumeSpecName: "config-data") pod "16151fa2-9a89-4a5e-a1c7-4853427e8851" (UID: "16151fa2-9a89-4a5e-a1c7-4853427e8851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.855082 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16151fa2-9a89-4a5e-a1c7-4853427e8851" (UID: "16151fa2-9a89-4a5e-a1c7-4853427e8851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.926620 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxfgw\" (UniqueName: \"kubernetes.io/projected/16151fa2-9a89-4a5e-a1c7-4853427e8851-kube-api-access-wxfgw\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.926717 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:39 crc kubenswrapper[4788]: I1010 16:19:39.926742 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16151fa2-9a89-4a5e-a1c7-4853427e8851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.259108 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" path="/var/lib/kubelet/pods/8ea194fc-2dde-490a-b9d5-6f17dfd3af19/volumes" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.378057 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"16151fa2-9a89-4a5e-a1c7-4853427e8851","Type":"ContainerDied","Data":"c335ec8c4bc35793fc142444eccdf2d65dec60f7500f02b73f28c70900c8b8a7"} Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.380205 4788 scope.go:117] "RemoveContainer" containerID="9079b3ba03acb325eff3c901f40d1fb3168c67d2061b3cb665d12702d2494d96" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.378347 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.418667 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.449329 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.464572 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:40 crc kubenswrapper[4788]: E1010 16:19:40.465164 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" containerName="init" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.465190 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" containerName="init" Oct 10 16:19:40 crc kubenswrapper[4788]: E1010 16:19:40.465206 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" containerName="dnsmasq-dns" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.465216 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" containerName="dnsmasq-dns" Oct 10 16:19:40 crc kubenswrapper[4788]: E1010 16:19:40.465252 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16151fa2-9a89-4a5e-a1c7-4853427e8851" containerName="nova-scheduler-scheduler" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.465263 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="16151fa2-9a89-4a5e-a1c7-4853427e8851" containerName="nova-scheduler-scheduler" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.465537 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea194fc-2dde-490a-b9d5-6f17dfd3af19" containerName="dnsmasq-dns" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.465585 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="16151fa2-9a89-4a5e-a1c7-4853427e8851" containerName="nova-scheduler-scheduler" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.466686 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.475527 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.484528 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.642521 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khj5s\" (UniqueName: \"kubernetes.io/projected/d04b7ef7-e70f-4471-bb15-b11fb1352120-kube-api-access-khj5s\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.642612 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.642673 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-config-data\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.745690 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-config-data\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.745934 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khj5s\" (UniqueName: \"kubernetes.io/projected/d04b7ef7-e70f-4471-bb15-b11fb1352120-kube-api-access-khj5s\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.746047 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.753420 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.757322 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-config-data\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.768899 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khj5s\" (UniqueName: \"kubernetes.io/projected/d04b7ef7-e70f-4471-bb15-b11fb1352120-kube-api-access-khj5s\") pod \"nova-scheduler-0\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:40 crc kubenswrapper[4788]: I1010 16:19:40.797629 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:41 crc kubenswrapper[4788]: I1010 16:19:41.308551 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:41 crc kubenswrapper[4788]: W1010 16:19:41.316758 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd04b7ef7_e70f_4471_bb15_b11fb1352120.slice/crio-7faaa0e040637a942735a2d1929b75e0b456adefd27895818ad592a914a229a0 WatchSource:0}: Error finding container 7faaa0e040637a942735a2d1929b75e0b456adefd27895818ad592a914a229a0: Status 404 returned error can't find the container with id 7faaa0e040637a942735a2d1929b75e0b456adefd27895818ad592a914a229a0 Oct 10 16:19:41 crc kubenswrapper[4788]: I1010 16:19:41.394551 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d04b7ef7-e70f-4471-bb15-b11fb1352120","Type":"ContainerStarted","Data":"7faaa0e040637a942735a2d1929b75e0b456adefd27895818ad592a914a229a0"} Oct 10 16:19:42 crc kubenswrapper[4788]: I1010 16:19:42.112838 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 16:19:42 crc kubenswrapper[4788]: I1010 16:19:42.112912 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 16:19:42 crc kubenswrapper[4788]: I1010 16:19:42.257270 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16151fa2-9a89-4a5e-a1c7-4853427e8851" path="/var/lib/kubelet/pods/16151fa2-9a89-4a5e-a1c7-4853427e8851/volumes" Oct 10 16:19:42 crc kubenswrapper[4788]: I1010 16:19:42.411831 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d04b7ef7-e70f-4471-bb15-b11fb1352120","Type":"ContainerStarted","Data":"fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec"} Oct 10 16:19:42 crc kubenswrapper[4788]: I1010 16:19:42.444822 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.444801032 podStartE2EDuration="2.444801032s" podCreationTimestamp="2025-10-10 16:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:42.431282474 +0000 UTC m=+5684.880998032" watchObservedRunningTime="2025-10-10 16:19:42.444801032 +0000 UTC m=+5684.894516590" Oct 10 16:19:44 crc kubenswrapper[4788]: I1010 16:19:44.690897 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.271914 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-6j2kj"] Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.273596 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.276677 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.276815 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.284278 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6j2kj"] Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.445088 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9lrt\" (UniqueName: \"kubernetes.io/projected/3021cf2d-3cf9-42c8-8a45-52bf67618e27-kube-api-access-q9lrt\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.445743 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-scripts\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.445886 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.446062 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-config-data\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.549480 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-scripts\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.549569 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.549630 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-config-data\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.549770 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9lrt\" (UniqueName: \"kubernetes.io/projected/3021cf2d-3cf9-42c8-8a45-52bf67618e27-kube-api-access-q9lrt\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.563329 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.564643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-scripts\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.565866 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-config-data\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.581640 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9lrt\" (UniqueName: \"kubernetes.io/projected/3021cf2d-3cf9-42c8-8a45-52bf67618e27-kube-api-access-q9lrt\") pod \"nova-cell1-cell-mapping-6j2kj\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.657612 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:45 crc kubenswrapper[4788]: I1010 16:19:45.798401 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 16:19:46 crc kubenswrapper[4788]: I1010 16:19:46.167084 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6j2kj"] Oct 10 16:19:46 crc kubenswrapper[4788]: W1010 16:19:46.168329 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3021cf2d_3cf9_42c8_8a45_52bf67618e27.slice/crio-280506ffc34ae3170de82f2bcd52e36851890ebb91ec2af1af296ba72d84060e WatchSource:0}: Error finding container 280506ffc34ae3170de82f2bcd52e36851890ebb91ec2af1af296ba72d84060e: Status 404 returned error can't find the container with id 280506ffc34ae3170de82f2bcd52e36851890ebb91ec2af1af296ba72d84060e Oct 10 16:19:46 crc kubenswrapper[4788]: I1010 16:19:46.486494 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6j2kj" event={"ID":"3021cf2d-3cf9-42c8-8a45-52bf67618e27","Type":"ContainerStarted","Data":"9a75079f37187cb3ddd27b0c52c050d7a897a0542fd238a327dfd4517ee27cfa"} Oct 10 16:19:46 crc kubenswrapper[4788]: I1010 16:19:46.486561 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6j2kj" event={"ID":"3021cf2d-3cf9-42c8-8a45-52bf67618e27","Type":"ContainerStarted","Data":"280506ffc34ae3170de82f2bcd52e36851890ebb91ec2af1af296ba72d84060e"} Oct 10 16:19:46 crc kubenswrapper[4788]: I1010 16:19:46.512032 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-6j2kj" podStartSLOduration=1.512003337 podStartE2EDuration="1.512003337s" podCreationTimestamp="2025-10-10 16:19:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:46.501865347 +0000 UTC m=+5688.951580895" watchObservedRunningTime="2025-10-10 16:19:46.512003337 +0000 UTC m=+5688.961718895" Oct 10 16:19:47 crc kubenswrapper[4788]: I1010 16:19:47.075650 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 16:19:47 crc kubenswrapper[4788]: I1010 16:19:47.075722 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 16:19:47 crc kubenswrapper[4788]: I1010 16:19:47.113493 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 16:19:47 crc kubenswrapper[4788]: I1010 16:19:47.113876 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 16:19:48 crc kubenswrapper[4788]: I1010 16:19:48.159391 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.71:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:19:48 crc kubenswrapper[4788]: I1010 16:19:48.242053 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:19:48 crc kubenswrapper[4788]: E1010 16:19:48.242379 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:19:48 crc kubenswrapper[4788]: I1010 16:19:48.242388 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.71:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:19:48 crc kubenswrapper[4788]: I1010 16:19:48.242478 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.72:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:19:48 crc kubenswrapper[4788]: I1010 16:19:48.242830 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.72:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:19:50 crc kubenswrapper[4788]: I1010 16:19:50.798825 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 16:19:50 crc kubenswrapper[4788]: I1010 16:19:50.847439 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 16:19:51 crc kubenswrapper[4788]: I1010 16:19:51.554430 4788 generic.go:334] "Generic (PLEG): container finished" podID="3021cf2d-3cf9-42c8-8a45-52bf67618e27" containerID="9a75079f37187cb3ddd27b0c52c050d7a897a0542fd238a327dfd4517ee27cfa" exitCode=0 Oct 10 16:19:51 crc kubenswrapper[4788]: I1010 16:19:51.554536 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6j2kj" event={"ID":"3021cf2d-3cf9-42c8-8a45-52bf67618e27","Type":"ContainerDied","Data":"9a75079f37187cb3ddd27b0c52c050d7a897a0542fd238a327dfd4517ee27cfa"} Oct 10 16:19:51 crc kubenswrapper[4788]: I1010 16:19:51.589794 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 16:19:52 crc kubenswrapper[4788]: I1010 16:19:52.936240 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.105837 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9lrt\" (UniqueName: \"kubernetes.io/projected/3021cf2d-3cf9-42c8-8a45-52bf67618e27-kube-api-access-q9lrt\") pod \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.105934 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-combined-ca-bundle\") pod \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.106043 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-config-data\") pod \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.106081 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-scripts\") pod \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\" (UID: \"3021cf2d-3cf9-42c8-8a45-52bf67618e27\") " Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.112065 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-scripts" (OuterVolumeSpecName: "scripts") pod "3021cf2d-3cf9-42c8-8a45-52bf67618e27" (UID: "3021cf2d-3cf9-42c8-8a45-52bf67618e27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.112968 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3021cf2d-3cf9-42c8-8a45-52bf67618e27-kube-api-access-q9lrt" (OuterVolumeSpecName: "kube-api-access-q9lrt") pod "3021cf2d-3cf9-42c8-8a45-52bf67618e27" (UID: "3021cf2d-3cf9-42c8-8a45-52bf67618e27"). InnerVolumeSpecName "kube-api-access-q9lrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.135189 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-config-data" (OuterVolumeSpecName: "config-data") pod "3021cf2d-3cf9-42c8-8a45-52bf67618e27" (UID: "3021cf2d-3cf9-42c8-8a45-52bf67618e27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.138359 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3021cf2d-3cf9-42c8-8a45-52bf67618e27" (UID: "3021cf2d-3cf9-42c8-8a45-52bf67618e27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.208061 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.208095 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9lrt\" (UniqueName: \"kubernetes.io/projected/3021cf2d-3cf9-42c8-8a45-52bf67618e27-kube-api-access-q9lrt\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.208111 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.208123 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3021cf2d-3cf9-42c8-8a45-52bf67618e27-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.575125 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6j2kj" event={"ID":"3021cf2d-3cf9-42c8-8a45-52bf67618e27","Type":"ContainerDied","Data":"280506ffc34ae3170de82f2bcd52e36851890ebb91ec2af1af296ba72d84060e"} Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.575186 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="280506ffc34ae3170de82f2bcd52e36851890ebb91ec2af1af296ba72d84060e" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.575253 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6j2kj" Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.766982 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.767284 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-log" containerID="cri-o://ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1" gracePeriod=30 Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.767355 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-api" containerID="cri-o://97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1" gracePeriod=30 Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.778804 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.779032 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d04b7ef7-e70f-4471-bb15-b11fb1352120" containerName="nova-scheduler-scheduler" containerID="cri-o://fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec" gracePeriod=30 Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.809655 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.815653 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-metadata" containerID="cri-o://823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a" gracePeriod=30 Oct 10 16:19:53 crc kubenswrapper[4788]: I1010 16:19:53.815623 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-log" containerID="cri-o://b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f" gracePeriod=30 Oct 10 16:19:54 crc kubenswrapper[4788]: I1010 16:19:54.586710 4788 generic.go:334] "Generic (PLEG): container finished" podID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerID="ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1" exitCode=143 Oct 10 16:19:54 crc kubenswrapper[4788]: I1010 16:19:54.586790 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a531e3a-9184-4aa4-9a68-5857a2a3df19","Type":"ContainerDied","Data":"ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1"} Oct 10 16:19:54 crc kubenswrapper[4788]: I1010 16:19:54.589726 4788 generic.go:334] "Generic (PLEG): container finished" podID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerID="b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f" exitCode=143 Oct 10 16:19:54 crc kubenswrapper[4788]: I1010 16:19:54.589764 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"944a91e4-1502-4abc-a51d-7fe2b61cb635","Type":"ContainerDied","Data":"b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f"} Oct 10 16:19:55 crc kubenswrapper[4788]: E1010 16:19:55.800777 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 16:19:55 crc kubenswrapper[4788]: E1010 16:19:55.803824 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 16:19:55 crc kubenswrapper[4788]: E1010 16:19:55.806072 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 10 16:19:55 crc kubenswrapper[4788]: E1010 16:19:55.806163 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d04b7ef7-e70f-4471-bb15-b11fb1352120" containerName="nova-scheduler-scheduler" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.040762 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a531e3a_9184_4aa4_9a68_5857a2a3df19.slice/crio-97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a531e3a_9184_4aa4_9a68_5857a2a3df19.slice/crio-conmon-97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1.scope\": RecentStats: unable to find data in memory cache]" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.449044 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.458081 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.583863 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xp5p\" (UniqueName: \"kubernetes.io/projected/944a91e4-1502-4abc-a51d-7fe2b61cb635-kube-api-access-8xp5p\") pod \"944a91e4-1502-4abc-a51d-7fe2b61cb635\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.584006 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a531e3a-9184-4aa4-9a68-5857a2a3df19-logs\") pod \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.584077 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-combined-ca-bundle\") pod \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.584104 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/944a91e4-1502-4abc-a51d-7fe2b61cb635-logs\") pod \"944a91e4-1502-4abc-a51d-7fe2b61cb635\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.584125 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-config-data\") pod \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.584207 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pgvk\" (UniqueName: \"kubernetes.io/projected/0a531e3a-9184-4aa4-9a68-5857a2a3df19-kube-api-access-2pgvk\") pod \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\" (UID: \"0a531e3a-9184-4aa4-9a68-5857a2a3df19\") " Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.584298 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-combined-ca-bundle\") pod \"944a91e4-1502-4abc-a51d-7fe2b61cb635\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.584362 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-config-data\") pod \"944a91e4-1502-4abc-a51d-7fe2b61cb635\" (UID: \"944a91e4-1502-4abc-a51d-7fe2b61cb635\") " Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.585687 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/944a91e4-1502-4abc-a51d-7fe2b61cb635-logs" (OuterVolumeSpecName: "logs") pod "944a91e4-1502-4abc-a51d-7fe2b61cb635" (UID: "944a91e4-1502-4abc-a51d-7fe2b61cb635"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.586303 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a531e3a-9184-4aa4-9a68-5857a2a3df19-logs" (OuterVolumeSpecName: "logs") pod "0a531e3a-9184-4aa4-9a68-5857a2a3df19" (UID: "0a531e3a-9184-4aa4-9a68-5857a2a3df19"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.589449 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a531e3a-9184-4aa4-9a68-5857a2a3df19-kube-api-access-2pgvk" (OuterVolumeSpecName: "kube-api-access-2pgvk") pod "0a531e3a-9184-4aa4-9a68-5857a2a3df19" (UID: "0a531e3a-9184-4aa4-9a68-5857a2a3df19"). InnerVolumeSpecName "kube-api-access-2pgvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.591438 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944a91e4-1502-4abc-a51d-7fe2b61cb635-kube-api-access-8xp5p" (OuterVolumeSpecName: "kube-api-access-8xp5p") pod "944a91e4-1502-4abc-a51d-7fe2b61cb635" (UID: "944a91e4-1502-4abc-a51d-7fe2b61cb635"). InnerVolumeSpecName "kube-api-access-8xp5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.613790 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-config-data" (OuterVolumeSpecName: "config-data") pod "944a91e4-1502-4abc-a51d-7fe2b61cb635" (UID: "944a91e4-1502-4abc-a51d-7fe2b61cb635"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.620060 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a531e3a-9184-4aa4-9a68-5857a2a3df19" (UID: "0a531e3a-9184-4aa4-9a68-5857a2a3df19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.621860 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-config-data" (OuterVolumeSpecName: "config-data") pod "0a531e3a-9184-4aa4-9a68-5857a2a3df19" (UID: "0a531e3a-9184-4aa4-9a68-5857a2a3df19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.622837 4788 generic.go:334] "Generic (PLEG): container finished" podID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerID="823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a" exitCode=0 Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.622909 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.622948 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"944a91e4-1502-4abc-a51d-7fe2b61cb635","Type":"ContainerDied","Data":"823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a"} Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.623000 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"944a91e4-1502-4abc-a51d-7fe2b61cb635","Type":"ContainerDied","Data":"c93b6d04be9d870eb78885e36f50060c8e525e2a46ac202de0319dca96d31073"} Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.623023 4788 scope.go:117] "RemoveContainer" containerID="823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.625643 4788 generic.go:334] "Generic (PLEG): container finished" podID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerID="97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1" exitCode=0 Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.625691 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a531e3a-9184-4aa4-9a68-5857a2a3df19","Type":"ContainerDied","Data":"97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1"} Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.625720 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a531e3a-9184-4aa4-9a68-5857a2a3df19","Type":"ContainerDied","Data":"e4720ccbc105b07726d490fbad296b36f2cb29e91cbd8d2ee07af7ea7845033c"} Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.625784 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.643996 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "944a91e4-1502-4abc-a51d-7fe2b61cb635" (UID: "944a91e4-1502-4abc-a51d-7fe2b61cb635"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.686379 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.686416 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944a91e4-1502-4abc-a51d-7fe2b61cb635-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.686425 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xp5p\" (UniqueName: \"kubernetes.io/projected/944a91e4-1502-4abc-a51d-7fe2b61cb635-kube-api-access-8xp5p\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.686435 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a531e3a-9184-4aa4-9a68-5857a2a3df19-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.686444 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.686453 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/944a91e4-1502-4abc-a51d-7fe2b61cb635-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.686462 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a531e3a-9184-4aa4-9a68-5857a2a3df19-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.686470 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pgvk\" (UniqueName: \"kubernetes.io/projected/0a531e3a-9184-4aa4-9a68-5857a2a3df19-kube-api-access-2pgvk\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.695856 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.703986 4788 scope.go:117] "RemoveContainer" containerID="b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.715989 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.728404 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.728852 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-log" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.728870 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-log" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.728895 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-metadata" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.728902 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-metadata" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.728923 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3021cf2d-3cf9-42c8-8a45-52bf67618e27" containerName="nova-manage" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.728929 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3021cf2d-3cf9-42c8-8a45-52bf67618e27" containerName="nova-manage" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.728938 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-api" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.728945 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-api" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.728956 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-log" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.728961 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-log" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.729136 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3021cf2d-3cf9-42c8-8a45-52bf67618e27" containerName="nova-manage" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.729193 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-api" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.729201 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" containerName="nova-api-log" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.729214 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-log" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.729224 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" containerName="nova-metadata-metadata" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.730133 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.732010 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.736314 4788 scope.go:117] "RemoveContainer" containerID="823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.738092 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a\": container with ID starting with 823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a not found: ID does not exist" containerID="823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.738124 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a"} err="failed to get container status \"823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a\": rpc error: code = NotFound desc = could not find container \"823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a\": container with ID starting with 823b107386aa3d96904778c90169e6a0f872786b9618b412d559e26998ae300a not found: ID does not exist" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.738167 4788 scope.go:117] "RemoveContainer" containerID="b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.738337 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f\": container with ID starting with b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f not found: ID does not exist" containerID="b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.738435 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f"} err="failed to get container status \"b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f\": rpc error: code = NotFound desc = could not find container \"b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f\": container with ID starting with b0113f8a4faed9316a4aeeeb39978b479a7a0edd8b695bfd7eb66593d7fb694f not found: ID does not exist" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.738448 4788 scope.go:117] "RemoveContainer" containerID="97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.743365 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.767857 4788 scope.go:117] "RemoveContainer" containerID="ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.784729 4788 scope.go:117] "RemoveContainer" containerID="97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.785259 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1\": container with ID starting with 97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1 not found: ID does not exist" containerID="97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.785313 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1"} err="failed to get container status \"97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1\": rpc error: code = NotFound desc = could not find container \"97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1\": container with ID starting with 97ca949295f46e7a8a505ccd46a1bdbd1f2327419230e7802273029b11ddbcf1 not found: ID does not exist" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.785339 4788 scope.go:117] "RemoveContainer" containerID="ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1" Oct 10 16:19:57 crc kubenswrapper[4788]: E1010 16:19:57.785734 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1\": container with ID starting with ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1 not found: ID does not exist" containerID="ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.785769 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1"} err="failed to get container status \"ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1\": rpc error: code = NotFound desc = could not find container \"ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1\": container with ID starting with ca7ff579c162759a25f98dc0caba2d147e26cc16181bdc1ff66d412ea17214f1 not found: ID does not exist" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.889589 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.889677 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4809b3c-2094-4a75-b601-792ebe980c9b-logs\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.889725 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-config-data\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.889776 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbn6z\" (UniqueName: \"kubernetes.io/projected/d4809b3c-2094-4a75-b601-792ebe980c9b-kube-api-access-bbn6z\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.970276 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.994248 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.996215 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:57 crc kubenswrapper[4788]: I1010 16:19:57.998249 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.002735 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.004973 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4809b3c-2094-4a75-b601-792ebe980c9b-logs\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.005053 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-config-data\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.005119 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbn6z\" (UniqueName: \"kubernetes.io/projected/d4809b3c-2094-4a75-b601-792ebe980c9b-kube-api-access-bbn6z\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.005214 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.009811 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.011438 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4809b3c-2094-4a75-b601-792ebe980c9b-logs\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.013101 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-config-data\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.014610 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.029848 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbn6z\" (UniqueName: \"kubernetes.io/projected/d4809b3c-2094-4a75-b601-792ebe980c9b-kube-api-access-bbn6z\") pod \"nova-api-0\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.056060 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.106635 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxjzr\" (UniqueName: \"kubernetes.io/projected/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-kube-api-access-bxjzr\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.106695 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.106738 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-config-data\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.106814 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-logs\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.204722 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.208552 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxjzr\" (UniqueName: \"kubernetes.io/projected/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-kube-api-access-bxjzr\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.208602 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.208652 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-config-data\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.208706 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-logs\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.209192 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-logs\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.222775 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-config-data\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.225209 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.225320 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxjzr\" (UniqueName: \"kubernetes.io/projected/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-kube-api-access-bxjzr\") pod \"nova-metadata-0\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.250857 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a531e3a-9184-4aa4-9a68-5857a2a3df19" path="/var/lib/kubelet/pods/0a531e3a-9184-4aa4-9a68-5857a2a3df19/volumes" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.255214 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="944a91e4-1502-4abc-a51d-7fe2b61cb635" path="/var/lib/kubelet/pods/944a91e4-1502-4abc-a51d-7fe2b61cb635/volumes" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.319438 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-config-data\") pod \"d04b7ef7-e70f-4471-bb15-b11fb1352120\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.319671 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khj5s\" (UniqueName: \"kubernetes.io/projected/d04b7ef7-e70f-4471-bb15-b11fb1352120-kube-api-access-khj5s\") pod \"d04b7ef7-e70f-4471-bb15-b11fb1352120\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.319763 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-combined-ca-bundle\") pod \"d04b7ef7-e70f-4471-bb15-b11fb1352120\" (UID: \"d04b7ef7-e70f-4471-bb15-b11fb1352120\") " Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.323353 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04b7ef7-e70f-4471-bb15-b11fb1352120-kube-api-access-khj5s" (OuterVolumeSpecName: "kube-api-access-khj5s") pod "d04b7ef7-e70f-4471-bb15-b11fb1352120" (UID: "d04b7ef7-e70f-4471-bb15-b11fb1352120"). InnerVolumeSpecName "kube-api-access-khj5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.352736 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-config-data" (OuterVolumeSpecName: "config-data") pod "d04b7ef7-e70f-4471-bb15-b11fb1352120" (UID: "d04b7ef7-e70f-4471-bb15-b11fb1352120"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.360615 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d04b7ef7-e70f-4471-bb15-b11fb1352120" (UID: "d04b7ef7-e70f-4471-bb15-b11fb1352120"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.421656 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.421688 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khj5s\" (UniqueName: \"kubernetes.io/projected/d04b7ef7-e70f-4471-bb15-b11fb1352120-kube-api-access-khj5s\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.421699 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d04b7ef7-e70f-4471-bb15-b11fb1352120-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.495556 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.502028 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:19:58 crc kubenswrapper[4788]: W1010 16:19:58.503435 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4809b3c_2094_4a75_b601_792ebe980c9b.slice/crio-3edb03b1bead459b65cd53fc64c2b5e8667c78c2706a278f082324ea93fe2185 WatchSource:0}: Error finding container 3edb03b1bead459b65cd53fc64c2b5e8667c78c2706a278f082324ea93fe2185: Status 404 returned error can't find the container with id 3edb03b1bead459b65cd53fc64c2b5e8667c78c2706a278f082324ea93fe2185 Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.661307 4788 generic.go:334] "Generic (PLEG): container finished" podID="d04b7ef7-e70f-4471-bb15-b11fb1352120" containerID="fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec" exitCode=0 Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.661356 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d04b7ef7-e70f-4471-bb15-b11fb1352120","Type":"ContainerDied","Data":"fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec"} Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.661378 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d04b7ef7-e70f-4471-bb15-b11fb1352120","Type":"ContainerDied","Data":"7faaa0e040637a942735a2d1929b75e0b456adefd27895818ad592a914a229a0"} Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.661394 4788 scope.go:117] "RemoveContainer" containerID="fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.661523 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.664514 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4809b3c-2094-4a75-b601-792ebe980c9b","Type":"ContainerStarted","Data":"3edb03b1bead459b65cd53fc64c2b5e8667c78c2706a278f082324ea93fe2185"} Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.686398 4788 scope.go:117] "RemoveContainer" containerID="fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.710858 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:58 crc kubenswrapper[4788]: E1010 16:19:58.712310 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec\": container with ID starting with fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec not found: ID does not exist" containerID="fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.712349 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec"} err="failed to get container status \"fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec\": rpc error: code = NotFound desc = could not find container \"fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec\": container with ID starting with fa98daecd4cae9b68de188175b870b75630ec2dd9e6a0f419369dfca77bfcdec not found: ID does not exist" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.721763 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.732343 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:58 crc kubenswrapper[4788]: E1010 16:19:58.732738 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04b7ef7-e70f-4471-bb15-b11fb1352120" containerName="nova-scheduler-scheduler" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.732761 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04b7ef7-e70f-4471-bb15-b11fb1352120" containerName="nova-scheduler-scheduler" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.732996 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04b7ef7-e70f-4471-bb15-b11fb1352120" containerName="nova-scheduler-scheduler" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.733760 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.735708 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.743872 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.861490 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58pnp\" (UniqueName: \"kubernetes.io/projected/7b94d62f-25ad-4615-8d4e-863e5362b66f-kube-api-access-58pnp\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.861527 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-config-data\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.861547 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.962893 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58pnp\" (UniqueName: \"kubernetes.io/projected/7b94d62f-25ad-4615-8d4e-863e5362b66f-kube-api-access-58pnp\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.962943 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-config-data\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.962975 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.966420 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-config-data\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.966447 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:58 crc kubenswrapper[4788]: I1010 16:19:58.978379 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58pnp\" (UniqueName: \"kubernetes.io/projected/7b94d62f-25ad-4615-8d4e-863e5362b66f-kube-api-access-58pnp\") pod \"nova-scheduler-0\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " pod="openstack/nova-scheduler-0" Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.020179 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:19:59 crc kubenswrapper[4788]: W1010 16:19:59.023363 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea232c34_bec6_4e90_a6fd_9a3ec7d520f4.slice/crio-cb075dea9e65dea2e1381a425f593ce2f9b29e0beedb3fdfcfac99363aa3e40b WatchSource:0}: Error finding container cb075dea9e65dea2e1381a425f593ce2f9b29e0beedb3fdfcfac99363aa3e40b: Status 404 returned error can't find the container with id cb075dea9e65dea2e1381a425f593ce2f9b29e0beedb3fdfcfac99363aa3e40b Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.064288 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.234208 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:19:59 crc kubenswrapper[4788]: E1010 16:19:59.235017 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.485467 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:19:59 crc kubenswrapper[4788]: W1010 16:19:59.507002 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b94d62f_25ad_4615_8d4e_863e5362b66f.slice/crio-540b5be8f7ae77b694bf8f51ac6433dd21907318afca32c8b3db8c116ec2e9f5 WatchSource:0}: Error finding container 540b5be8f7ae77b694bf8f51ac6433dd21907318afca32c8b3db8c116ec2e9f5: Status 404 returned error can't find the container with id 540b5be8f7ae77b694bf8f51ac6433dd21907318afca32c8b3db8c116ec2e9f5 Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.686228 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4809b3c-2094-4a75-b601-792ebe980c9b","Type":"ContainerStarted","Data":"5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2"} Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.686302 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4809b3c-2094-4a75-b601-792ebe980c9b","Type":"ContainerStarted","Data":"3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb"} Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.690629 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4","Type":"ContainerStarted","Data":"960c4c7bf19c6b08d2b63577629db693d8befa34eead86a05bd574119e876057"} Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.690682 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4","Type":"ContainerStarted","Data":"ff8ba4238c632907444b0341c93de733d6cd96167489db240b70f3285ac0b32f"} Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.690700 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4","Type":"ContainerStarted","Data":"cb075dea9e65dea2e1381a425f593ce2f9b29e0beedb3fdfcfac99363aa3e40b"} Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.693619 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7b94d62f-25ad-4615-8d4e-863e5362b66f","Type":"ContainerStarted","Data":"540b5be8f7ae77b694bf8f51ac6433dd21907318afca32c8b3db8c116ec2e9f5"} Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.731010 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.73098884 podStartE2EDuration="2.73098884s" podCreationTimestamp="2025-10-10 16:19:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:59.709074807 +0000 UTC m=+5702.158790375" watchObservedRunningTime="2025-10-10 16:19:59.73098884 +0000 UTC m=+5702.180704398" Oct 10 16:19:59 crc kubenswrapper[4788]: I1010 16:19:59.733307 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7332924690000002 podStartE2EDuration="2.733292469s" podCreationTimestamp="2025-10-10 16:19:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:19:59.729485811 +0000 UTC m=+5702.179201369" watchObservedRunningTime="2025-10-10 16:19:59.733292469 +0000 UTC m=+5702.183008017" Oct 10 16:20:00 crc kubenswrapper[4788]: I1010 16:20:00.248011 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d04b7ef7-e70f-4471-bb15-b11fb1352120" path="/var/lib/kubelet/pods/d04b7ef7-e70f-4471-bb15-b11fb1352120/volumes" Oct 10 16:20:00 crc kubenswrapper[4788]: I1010 16:20:00.706610 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7b94d62f-25ad-4615-8d4e-863e5362b66f","Type":"ContainerStarted","Data":"b54ac40c052f0988b2e86ceb19ce6aab75391b5aa327c8fc731086fd2f3239ab"} Oct 10 16:20:00 crc kubenswrapper[4788]: I1010 16:20:00.734213 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.734189332 podStartE2EDuration="2.734189332s" podCreationTimestamp="2025-10-10 16:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:20:00.72555336 +0000 UTC m=+5703.175268938" watchObservedRunningTime="2025-10-10 16:20:00.734189332 +0000 UTC m=+5703.183904880" Oct 10 16:20:03 crc kubenswrapper[4788]: I1010 16:20:03.503388 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 16:20:03 crc kubenswrapper[4788]: I1010 16:20:03.504697 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 16:20:04 crc kubenswrapper[4788]: I1010 16:20:04.064697 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 16:20:08 crc kubenswrapper[4788]: I1010 16:20:08.057177 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 16:20:08 crc kubenswrapper[4788]: I1010 16:20:08.057992 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 16:20:08 crc kubenswrapper[4788]: I1010 16:20:08.503021 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 16:20:08 crc kubenswrapper[4788]: I1010 16:20:08.503092 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 16:20:09 crc kubenswrapper[4788]: I1010 16:20:09.064974 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 16:20:09 crc kubenswrapper[4788]: I1010 16:20:09.092648 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 16:20:09 crc kubenswrapper[4788]: I1010 16:20:09.139380 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.75:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:20:09 crc kubenswrapper[4788]: I1010 16:20:09.139482 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.75:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:20:09 crc kubenswrapper[4788]: I1010 16:20:09.585364 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:20:09 crc kubenswrapper[4788]: I1010 16:20:09.585385 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:20:09 crc kubenswrapper[4788]: I1010 16:20:09.837349 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 16:20:12 crc kubenswrapper[4788]: I1010 16:20:12.234986 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:20:12 crc kubenswrapper[4788]: E1010 16:20:12.235543 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.060644 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.061541 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.061876 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.061912 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.065225 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.066096 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.278507 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6785b99c97-2j5bq"] Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.282446 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.295171 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6785b99c97-2j5bq"] Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.349382 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82ddn\" (UniqueName: \"kubernetes.io/projected/ef80c08c-eef4-4e88-b18e-e43a38285f93-kube-api-access-82ddn\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.350420 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-dns-svc\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.350711 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-nb\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.351038 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-sb\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.351096 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-config\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.452877 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-sb\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.453165 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-config\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.453296 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82ddn\" (UniqueName: \"kubernetes.io/projected/ef80c08c-eef4-4e88-b18e-e43a38285f93-kube-api-access-82ddn\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.453457 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-dns-svc\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.453559 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-nb\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.454638 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-nb\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.455358 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-sb\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.456093 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-config\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.456321 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-dns-svc\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.480304 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82ddn\" (UniqueName: \"kubernetes.io/projected/ef80c08c-eef4-4e88-b18e-e43a38285f93-kube-api-access-82ddn\") pod \"dnsmasq-dns-6785b99c97-2j5bq\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.506117 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.508550 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.508998 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.605197 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:18 crc kubenswrapper[4788]: I1010 16:20:18.892746 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 16:20:19 crc kubenswrapper[4788]: I1010 16:20:19.112889 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6785b99c97-2j5bq"] Oct 10 16:20:19 crc kubenswrapper[4788]: W1010 16:20:19.117314 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef80c08c_eef4_4e88_b18e_e43a38285f93.slice/crio-b58ef9d21f2caacf8ea69a81d35e1ca43df6ecf17851eea14615871698ccec28 WatchSource:0}: Error finding container b58ef9d21f2caacf8ea69a81d35e1ca43df6ecf17851eea14615871698ccec28: Status 404 returned error can't find the container with id b58ef9d21f2caacf8ea69a81d35e1ca43df6ecf17851eea14615871698ccec28 Oct 10 16:20:19 crc kubenswrapper[4788]: I1010 16:20:19.901554 4788 generic.go:334] "Generic (PLEG): container finished" podID="ef80c08c-eef4-4e88-b18e-e43a38285f93" containerID="ec7c7d6a469c2235e28569f122029dc36db8dc5fc4fd1b59cc882184cbd91206" exitCode=0 Oct 10 16:20:19 crc kubenswrapper[4788]: I1010 16:20:19.903739 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" event={"ID":"ef80c08c-eef4-4e88-b18e-e43a38285f93","Type":"ContainerDied","Data":"ec7c7d6a469c2235e28569f122029dc36db8dc5fc4fd1b59cc882184cbd91206"} Oct 10 16:20:19 crc kubenswrapper[4788]: I1010 16:20:19.903784 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" event={"ID":"ef80c08c-eef4-4e88-b18e-e43a38285f93","Type":"ContainerStarted","Data":"b58ef9d21f2caacf8ea69a81d35e1ca43df6ecf17851eea14615871698ccec28"} Oct 10 16:20:20 crc kubenswrapper[4788]: I1010 16:20:20.912825 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" event={"ID":"ef80c08c-eef4-4e88-b18e-e43a38285f93","Type":"ContainerStarted","Data":"18d44731daaac1d4886a45be352fdbe6636494e040147ea6e69da5b55759a4a4"} Oct 10 16:20:20 crc kubenswrapper[4788]: I1010 16:20:20.936427 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" podStartSLOduration=2.936404814 podStartE2EDuration="2.936404814s" podCreationTimestamp="2025-10-10 16:20:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:20:20.927524975 +0000 UTC m=+5723.377240523" watchObservedRunningTime="2025-10-10 16:20:20.936404814 +0000 UTC m=+5723.386120352" Oct 10 16:20:21 crc kubenswrapper[4788]: I1010 16:20:21.922718 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:27 crc kubenswrapper[4788]: I1010 16:20:27.233398 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:20:27 crc kubenswrapper[4788]: E1010 16:20:27.234177 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:20:28 crc kubenswrapper[4788]: I1010 16:20:28.606302 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:20:28 crc kubenswrapper[4788]: I1010 16:20:28.678775 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fd4f44ddf-4qddw"] Oct 10 16:20:28 crc kubenswrapper[4788]: I1010 16:20:28.679212 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" podUID="ea15b81b-bb54-40bf-a470-70645b0e7759" containerName="dnsmasq-dns" containerID="cri-o://fd9d27226e6c45b71cbe8ed9cd691fa7779d8477be4919e335e7fafb8bb94218" gracePeriod=10 Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.007086 4788 generic.go:334] "Generic (PLEG): container finished" podID="ea15b81b-bb54-40bf-a470-70645b0e7759" containerID="fd9d27226e6c45b71cbe8ed9cd691fa7779d8477be4919e335e7fafb8bb94218" exitCode=0 Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.007135 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" event={"ID":"ea15b81b-bb54-40bf-a470-70645b0e7759","Type":"ContainerDied","Data":"fd9d27226e6c45b71cbe8ed9cd691fa7779d8477be4919e335e7fafb8bb94218"} Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.182284 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.274209 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-config\") pod \"ea15b81b-bb54-40bf-a470-70645b0e7759\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.274366 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bmph\" (UniqueName: \"kubernetes.io/projected/ea15b81b-bb54-40bf-a470-70645b0e7759-kube-api-access-6bmph\") pod \"ea15b81b-bb54-40bf-a470-70645b0e7759\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.274472 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-dns-svc\") pod \"ea15b81b-bb54-40bf-a470-70645b0e7759\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.274508 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-nb\") pod \"ea15b81b-bb54-40bf-a470-70645b0e7759\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.274553 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-sb\") pod \"ea15b81b-bb54-40bf-a470-70645b0e7759\" (UID: \"ea15b81b-bb54-40bf-a470-70645b0e7759\") " Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.281667 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea15b81b-bb54-40bf-a470-70645b0e7759-kube-api-access-6bmph" (OuterVolumeSpecName: "kube-api-access-6bmph") pod "ea15b81b-bb54-40bf-a470-70645b0e7759" (UID: "ea15b81b-bb54-40bf-a470-70645b0e7759"). InnerVolumeSpecName "kube-api-access-6bmph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.351600 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-config" (OuterVolumeSpecName: "config") pod "ea15b81b-bb54-40bf-a470-70645b0e7759" (UID: "ea15b81b-bb54-40bf-a470-70645b0e7759"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.366802 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea15b81b-bb54-40bf-a470-70645b0e7759" (UID: "ea15b81b-bb54-40bf-a470-70645b0e7759"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.367533 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea15b81b-bb54-40bf-a470-70645b0e7759" (UID: "ea15b81b-bb54-40bf-a470-70645b0e7759"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.373494 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea15b81b-bb54-40bf-a470-70645b0e7759" (UID: "ea15b81b-bb54-40bf-a470-70645b0e7759"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.377598 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bmph\" (UniqueName: \"kubernetes.io/projected/ea15b81b-bb54-40bf-a470-70645b0e7759-kube-api-access-6bmph\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.377625 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.377637 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.377647 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:29 crc kubenswrapper[4788]: I1010 16:20:29.377657 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea15b81b-bb54-40bf-a470-70645b0e7759-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:30 crc kubenswrapper[4788]: I1010 16:20:30.031970 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" event={"ID":"ea15b81b-bb54-40bf-a470-70645b0e7759","Type":"ContainerDied","Data":"4af723edf17f9055636caf0225801764f3c3a3a9bb7a206d235a17c11372b21e"} Oct 10 16:20:30 crc kubenswrapper[4788]: I1010 16:20:30.032023 4788 scope.go:117] "RemoveContainer" containerID="fd9d27226e6c45b71cbe8ed9cd691fa7779d8477be4919e335e7fafb8bb94218" Oct 10 16:20:30 crc kubenswrapper[4788]: I1010 16:20:30.032127 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fd4f44ddf-4qddw" Oct 10 16:20:30 crc kubenswrapper[4788]: I1010 16:20:30.068800 4788 scope.go:117] "RemoveContainer" containerID="a713fb395ae24a60f376ea4be0235b6027261ee459409d2bb69aa57af3d2ede5" Oct 10 16:20:30 crc kubenswrapper[4788]: I1010 16:20:30.077748 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fd4f44ddf-4qddw"] Oct 10 16:20:30 crc kubenswrapper[4788]: I1010 16:20:30.087692 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fd4f44ddf-4qddw"] Oct 10 16:20:30 crc kubenswrapper[4788]: I1010 16:20:30.243512 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea15b81b-bb54-40bf-a470-70645b0e7759" path="/var/lib/kubelet/pods/ea15b81b-bb54-40bf-a470-70645b0e7759/volumes" Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.732385 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-cdcgn"] Oct 10 16:20:31 crc kubenswrapper[4788]: E1010 16:20:31.732826 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea15b81b-bb54-40bf-a470-70645b0e7759" containerName="dnsmasq-dns" Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.732840 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea15b81b-bb54-40bf-a470-70645b0e7759" containerName="dnsmasq-dns" Oct 10 16:20:31 crc kubenswrapper[4788]: E1010 16:20:31.732858 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea15b81b-bb54-40bf-a470-70645b0e7759" containerName="init" Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.732865 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea15b81b-bb54-40bf-a470-70645b0e7759" containerName="init" Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.733048 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea15b81b-bb54-40bf-a470-70645b0e7759" containerName="dnsmasq-dns" Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.733648 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cdcgn" Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.746187 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cdcgn"] Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.818081 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pg52\" (UniqueName: \"kubernetes.io/projected/de9be368-d869-476f-a24e-682a3ec8821d-kube-api-access-4pg52\") pod \"cinder-db-create-cdcgn\" (UID: \"de9be368-d869-476f-a24e-682a3ec8821d\") " pod="openstack/cinder-db-create-cdcgn" Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.920185 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pg52\" (UniqueName: \"kubernetes.io/projected/de9be368-d869-476f-a24e-682a3ec8821d-kube-api-access-4pg52\") pod \"cinder-db-create-cdcgn\" (UID: \"de9be368-d869-476f-a24e-682a3ec8821d\") " pod="openstack/cinder-db-create-cdcgn" Oct 10 16:20:31 crc kubenswrapper[4788]: I1010 16:20:31.942627 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pg52\" (UniqueName: \"kubernetes.io/projected/de9be368-d869-476f-a24e-682a3ec8821d-kube-api-access-4pg52\") pod \"cinder-db-create-cdcgn\" (UID: \"de9be368-d869-476f-a24e-682a3ec8821d\") " pod="openstack/cinder-db-create-cdcgn" Oct 10 16:20:32 crc kubenswrapper[4788]: I1010 16:20:32.089780 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cdcgn" Oct 10 16:20:32 crc kubenswrapper[4788]: I1010 16:20:32.528237 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cdcgn"] Oct 10 16:20:33 crc kubenswrapper[4788]: I1010 16:20:33.075791 4788 generic.go:334] "Generic (PLEG): container finished" podID="de9be368-d869-476f-a24e-682a3ec8821d" containerID="2aeb26981bc33b85edc78f4b183b3255c039e4977d876d4c4d00fbc1d3f25b1b" exitCode=0 Oct 10 16:20:33 crc kubenswrapper[4788]: I1010 16:20:33.075847 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cdcgn" event={"ID":"de9be368-d869-476f-a24e-682a3ec8821d","Type":"ContainerDied","Data":"2aeb26981bc33b85edc78f4b183b3255c039e4977d876d4c4d00fbc1d3f25b1b"} Oct 10 16:20:33 crc kubenswrapper[4788]: I1010 16:20:33.075916 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cdcgn" event={"ID":"de9be368-d869-476f-a24e-682a3ec8821d","Type":"ContainerStarted","Data":"9e145945883d48718106a64baaf403d64a3eb0de6c74b816efbe5eddb4a7cd45"} Oct 10 16:20:34 crc kubenswrapper[4788]: I1010 16:20:34.454050 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cdcgn" Oct 10 16:20:34 crc kubenswrapper[4788]: I1010 16:20:34.568404 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pg52\" (UniqueName: \"kubernetes.io/projected/de9be368-d869-476f-a24e-682a3ec8821d-kube-api-access-4pg52\") pod \"de9be368-d869-476f-a24e-682a3ec8821d\" (UID: \"de9be368-d869-476f-a24e-682a3ec8821d\") " Oct 10 16:20:34 crc kubenswrapper[4788]: I1010 16:20:34.578327 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9be368-d869-476f-a24e-682a3ec8821d-kube-api-access-4pg52" (OuterVolumeSpecName: "kube-api-access-4pg52") pod "de9be368-d869-476f-a24e-682a3ec8821d" (UID: "de9be368-d869-476f-a24e-682a3ec8821d"). InnerVolumeSpecName "kube-api-access-4pg52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:20:34 crc kubenswrapper[4788]: I1010 16:20:34.672202 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pg52\" (UniqueName: \"kubernetes.io/projected/de9be368-d869-476f-a24e-682a3ec8821d-kube-api-access-4pg52\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:35 crc kubenswrapper[4788]: I1010 16:20:35.100899 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cdcgn" event={"ID":"de9be368-d869-476f-a24e-682a3ec8821d","Type":"ContainerDied","Data":"9e145945883d48718106a64baaf403d64a3eb0de6c74b816efbe5eddb4a7cd45"} Oct 10 16:20:35 crc kubenswrapper[4788]: I1010 16:20:35.100995 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e145945883d48718106a64baaf403d64a3eb0de6c74b816efbe5eddb4a7cd45" Oct 10 16:20:35 crc kubenswrapper[4788]: I1010 16:20:35.100989 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cdcgn" Oct 10 16:20:41 crc kubenswrapper[4788]: I1010 16:20:41.865523 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-092d-account-create-dtpnc"] Oct 10 16:20:41 crc kubenswrapper[4788]: E1010 16:20:41.867446 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9be368-d869-476f-a24e-682a3ec8821d" containerName="mariadb-database-create" Oct 10 16:20:41 crc kubenswrapper[4788]: I1010 16:20:41.867472 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9be368-d869-476f-a24e-682a3ec8821d" containerName="mariadb-database-create" Oct 10 16:20:41 crc kubenswrapper[4788]: I1010 16:20:41.868945 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9be368-d869-476f-a24e-682a3ec8821d" containerName="mariadb-database-create" Oct 10 16:20:41 crc kubenswrapper[4788]: I1010 16:20:41.870294 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-092d-account-create-dtpnc" Oct 10 16:20:41 crc kubenswrapper[4788]: I1010 16:20:41.873961 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 10 16:20:41 crc kubenswrapper[4788]: I1010 16:20:41.890946 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-092d-account-create-dtpnc"] Oct 10 16:20:41 crc kubenswrapper[4788]: I1010 16:20:41.917067 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdx6d\" (UniqueName: \"kubernetes.io/projected/35d3d68a-818c-48af-bc70-09c7d11a9177-kube-api-access-bdx6d\") pod \"cinder-092d-account-create-dtpnc\" (UID: \"35d3d68a-818c-48af-bc70-09c7d11a9177\") " pod="openstack/cinder-092d-account-create-dtpnc" Oct 10 16:20:42 crc kubenswrapper[4788]: I1010 16:20:42.018626 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdx6d\" (UniqueName: \"kubernetes.io/projected/35d3d68a-818c-48af-bc70-09c7d11a9177-kube-api-access-bdx6d\") pod \"cinder-092d-account-create-dtpnc\" (UID: \"35d3d68a-818c-48af-bc70-09c7d11a9177\") " pod="openstack/cinder-092d-account-create-dtpnc" Oct 10 16:20:42 crc kubenswrapper[4788]: I1010 16:20:42.040201 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdx6d\" (UniqueName: \"kubernetes.io/projected/35d3d68a-818c-48af-bc70-09c7d11a9177-kube-api-access-bdx6d\") pod \"cinder-092d-account-create-dtpnc\" (UID: \"35d3d68a-818c-48af-bc70-09c7d11a9177\") " pod="openstack/cinder-092d-account-create-dtpnc" Oct 10 16:20:42 crc kubenswrapper[4788]: I1010 16:20:42.194986 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-092d-account-create-dtpnc" Oct 10 16:20:42 crc kubenswrapper[4788]: I1010 16:20:42.236064 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:20:42 crc kubenswrapper[4788]: E1010 16:20:42.236407 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:20:42 crc kubenswrapper[4788]: I1010 16:20:42.634741 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-092d-account-create-dtpnc"] Oct 10 16:20:43 crc kubenswrapper[4788]: I1010 16:20:43.173689 4788 generic.go:334] "Generic (PLEG): container finished" podID="35d3d68a-818c-48af-bc70-09c7d11a9177" containerID="c3d1eb399feb77ea0c3badde3438508613091fcaac460dbea92e86d2d067e614" exitCode=0 Oct 10 16:20:43 crc kubenswrapper[4788]: I1010 16:20:43.173892 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-092d-account-create-dtpnc" event={"ID":"35d3d68a-818c-48af-bc70-09c7d11a9177","Type":"ContainerDied","Data":"c3d1eb399feb77ea0c3badde3438508613091fcaac460dbea92e86d2d067e614"} Oct 10 16:20:43 crc kubenswrapper[4788]: I1010 16:20:43.174050 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-092d-account-create-dtpnc" event={"ID":"35d3d68a-818c-48af-bc70-09c7d11a9177","Type":"ContainerStarted","Data":"fe2fe2fc58a0d373113d3a1ad1613d7696cd32dc21a0d2b7239e57aced35d75e"} Oct 10 16:20:44 crc kubenswrapper[4788]: I1010 16:20:44.697279 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-092d-account-create-dtpnc" Oct 10 16:20:44 crc kubenswrapper[4788]: I1010 16:20:44.771000 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdx6d\" (UniqueName: \"kubernetes.io/projected/35d3d68a-818c-48af-bc70-09c7d11a9177-kube-api-access-bdx6d\") pod \"35d3d68a-818c-48af-bc70-09c7d11a9177\" (UID: \"35d3d68a-818c-48af-bc70-09c7d11a9177\") " Oct 10 16:20:44 crc kubenswrapper[4788]: I1010 16:20:44.777717 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d3d68a-818c-48af-bc70-09c7d11a9177-kube-api-access-bdx6d" (OuterVolumeSpecName: "kube-api-access-bdx6d") pod "35d3d68a-818c-48af-bc70-09c7d11a9177" (UID: "35d3d68a-818c-48af-bc70-09c7d11a9177"). InnerVolumeSpecName "kube-api-access-bdx6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:20:44 crc kubenswrapper[4788]: I1010 16:20:44.873748 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdx6d\" (UniqueName: \"kubernetes.io/projected/35d3d68a-818c-48af-bc70-09c7d11a9177-kube-api-access-bdx6d\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:45 crc kubenswrapper[4788]: I1010 16:20:45.198736 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-092d-account-create-dtpnc" event={"ID":"35d3d68a-818c-48af-bc70-09c7d11a9177","Type":"ContainerDied","Data":"fe2fe2fc58a0d373113d3a1ad1613d7696cd32dc21a0d2b7239e57aced35d75e"} Oct 10 16:20:45 crc kubenswrapper[4788]: I1010 16:20:45.198789 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe2fe2fc58a0d373113d3a1ad1613d7696cd32dc21a0d2b7239e57aced35d75e" Oct 10 16:20:45 crc kubenswrapper[4788]: I1010 16:20:45.198886 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-092d-account-create-dtpnc" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.199781 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-257v4"] Oct 10 16:20:47 crc kubenswrapper[4788]: E1010 16:20:47.200579 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d3d68a-818c-48af-bc70-09c7d11a9177" containerName="mariadb-account-create" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.200601 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d3d68a-818c-48af-bc70-09c7d11a9177" containerName="mariadb-account-create" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.200837 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d3d68a-818c-48af-bc70-09c7d11a9177" containerName="mariadb-account-create" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.201689 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.205877 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.206117 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7rrkh" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.206308 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.212965 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-257v4"] Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.324011 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8473942-7ab7-4e67-bc42-01e6c06d088f-etc-machine-id\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.324101 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-config-data\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.324357 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrxpf\" (UniqueName: \"kubernetes.io/projected/a8473942-7ab7-4e67-bc42-01e6c06d088f-kube-api-access-vrxpf\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.324658 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-scripts\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.324727 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-combined-ca-bundle\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.325042 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-db-sync-config-data\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.426865 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrxpf\" (UniqueName: \"kubernetes.io/projected/a8473942-7ab7-4e67-bc42-01e6c06d088f-kube-api-access-vrxpf\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.426961 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-scripts\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.426993 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-combined-ca-bundle\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.427091 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-db-sync-config-data\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.427175 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8473942-7ab7-4e67-bc42-01e6c06d088f-etc-machine-id\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.427266 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-config-data\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.428856 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8473942-7ab7-4e67-bc42-01e6c06d088f-etc-machine-id\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.432358 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-scripts\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.432566 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-combined-ca-bundle\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.434783 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-config-data\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.441331 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-db-sync-config-data\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.457765 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrxpf\" (UniqueName: \"kubernetes.io/projected/a8473942-7ab7-4e67-bc42-01e6c06d088f-kube-api-access-vrxpf\") pod \"cinder-db-sync-257v4\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:47 crc kubenswrapper[4788]: I1010 16:20:47.541370 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:48 crc kubenswrapper[4788]: I1010 16:20:48.008660 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-257v4"] Oct 10 16:20:48 crc kubenswrapper[4788]: I1010 16:20:48.251638 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-257v4" event={"ID":"a8473942-7ab7-4e67-bc42-01e6c06d088f","Type":"ContainerStarted","Data":"de1c17ecf477d26dc7d9dfab93439a172e7e48f4461d8ae11b9ae9f615a9df6b"} Oct 10 16:20:49 crc kubenswrapper[4788]: I1010 16:20:49.251847 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-257v4" event={"ID":"a8473942-7ab7-4e67-bc42-01e6c06d088f","Type":"ContainerStarted","Data":"5d20d1006436fbf2b65d3bd508ba74c1355a9279ab8fe1284ee857ff1ec7cd7a"} Oct 10 16:20:49 crc kubenswrapper[4788]: I1010 16:20:49.277609 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-257v4" podStartSLOduration=2.27752507 podStartE2EDuration="2.27752507s" podCreationTimestamp="2025-10-10 16:20:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:20:49.276701309 +0000 UTC m=+5751.726416947" watchObservedRunningTime="2025-10-10 16:20:49.27752507 +0000 UTC m=+5751.727240628" Oct 10 16:20:51 crc kubenswrapper[4788]: I1010 16:20:51.270197 4788 generic.go:334] "Generic (PLEG): container finished" podID="a8473942-7ab7-4e67-bc42-01e6c06d088f" containerID="5d20d1006436fbf2b65d3bd508ba74c1355a9279ab8fe1284ee857ff1ec7cd7a" exitCode=0 Oct 10 16:20:51 crc kubenswrapper[4788]: I1010 16:20:51.270294 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-257v4" event={"ID":"a8473942-7ab7-4e67-bc42-01e6c06d088f","Type":"ContainerDied","Data":"5d20d1006436fbf2b65d3bd508ba74c1355a9279ab8fe1284ee857ff1ec7cd7a"} Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.632649 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.716701 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-scripts\") pod \"a8473942-7ab7-4e67-bc42-01e6c06d088f\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.716801 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-combined-ca-bundle\") pod \"a8473942-7ab7-4e67-bc42-01e6c06d088f\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.716863 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrxpf\" (UniqueName: \"kubernetes.io/projected/a8473942-7ab7-4e67-bc42-01e6c06d088f-kube-api-access-vrxpf\") pod \"a8473942-7ab7-4e67-bc42-01e6c06d088f\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.716911 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-config-data\") pod \"a8473942-7ab7-4e67-bc42-01e6c06d088f\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.716949 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-db-sync-config-data\") pod \"a8473942-7ab7-4e67-bc42-01e6c06d088f\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.717646 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8473942-7ab7-4e67-bc42-01e6c06d088f-etc-machine-id\") pod \"a8473942-7ab7-4e67-bc42-01e6c06d088f\" (UID: \"a8473942-7ab7-4e67-bc42-01e6c06d088f\") " Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.717767 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8473942-7ab7-4e67-bc42-01e6c06d088f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a8473942-7ab7-4e67-bc42-01e6c06d088f" (UID: "a8473942-7ab7-4e67-bc42-01e6c06d088f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.718105 4788 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8473942-7ab7-4e67-bc42-01e6c06d088f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.722449 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a8473942-7ab7-4e67-bc42-01e6c06d088f" (UID: "a8473942-7ab7-4e67-bc42-01e6c06d088f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.722745 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-scripts" (OuterVolumeSpecName: "scripts") pod "a8473942-7ab7-4e67-bc42-01e6c06d088f" (UID: "a8473942-7ab7-4e67-bc42-01e6c06d088f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.722787 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8473942-7ab7-4e67-bc42-01e6c06d088f-kube-api-access-vrxpf" (OuterVolumeSpecName: "kube-api-access-vrxpf") pod "a8473942-7ab7-4e67-bc42-01e6c06d088f" (UID: "a8473942-7ab7-4e67-bc42-01e6c06d088f"). InnerVolumeSpecName "kube-api-access-vrxpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.743933 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8473942-7ab7-4e67-bc42-01e6c06d088f" (UID: "a8473942-7ab7-4e67-bc42-01e6c06d088f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.771613 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-config-data" (OuterVolumeSpecName: "config-data") pod "a8473942-7ab7-4e67-bc42-01e6c06d088f" (UID: "a8473942-7ab7-4e67-bc42-01e6c06d088f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.819883 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrxpf\" (UniqueName: \"kubernetes.io/projected/a8473942-7ab7-4e67-bc42-01e6c06d088f-kube-api-access-vrxpf\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.819912 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.819921 4788 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.819930 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:52 crc kubenswrapper[4788]: I1010 16:20:52.819939 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8473942-7ab7-4e67-bc42-01e6c06d088f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.295106 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-257v4" event={"ID":"a8473942-7ab7-4e67-bc42-01e6c06d088f","Type":"ContainerDied","Data":"de1c17ecf477d26dc7d9dfab93439a172e7e48f4461d8ae11b9ae9f615a9df6b"} Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.295290 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-257v4" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.295557 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1c17ecf477d26dc7d9dfab93439a172e7e48f4461d8ae11b9ae9f615a9df6b" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.628104 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b757d79cc-8ltdz"] Oct 10 16:20:53 crc kubenswrapper[4788]: E1010 16:20:53.628528 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8473942-7ab7-4e67-bc42-01e6c06d088f" containerName="cinder-db-sync" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.628544 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8473942-7ab7-4e67-bc42-01e6c06d088f" containerName="cinder-db-sync" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.628697 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8473942-7ab7-4e67-bc42-01e6c06d088f" containerName="cinder-db-sync" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.629602 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.656971 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b757d79cc-8ltdz"] Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.736339 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-sb\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.736421 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-dns-svc\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.736489 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5whx\" (UniqueName: \"kubernetes.io/projected/e2664a06-b38b-49b9-8651-75098ac9d1f0-kube-api-access-l5whx\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.736610 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-config\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.736644 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-nb\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.803368 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.805526 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.815179 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.817834 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.818028 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.818054 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.818484 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7rrkh" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.838053 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-dns-svc\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.838151 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5whx\" (UniqueName: \"kubernetes.io/projected/e2664a06-b38b-49b9-8651-75098ac9d1f0-kube-api-access-l5whx\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.838176 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-config\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.838191 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-nb\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.838255 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-sb\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.839089 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-nb\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.839112 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-sb\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.839124 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-dns-svc\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.839232 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-config\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.856919 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5whx\" (UniqueName: \"kubernetes.io/projected/e2664a06-b38b-49b9-8651-75098ac9d1f0-kube-api-access-l5whx\") pod \"dnsmasq-dns-b757d79cc-8ltdz\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.939279 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/398b7cc2-ff9d-4967-ad41-d89a87f831fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.939426 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.939446 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfpcj\" (UniqueName: \"kubernetes.io/projected/398b7cc2-ff9d-4967-ad41-d89a87f831fe-kube-api-access-bfpcj\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.939460 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-scripts\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.940575 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/398b7cc2-ff9d-4967-ad41-d89a87f831fe-logs\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.940673 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.940802 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:53 crc kubenswrapper[4788]: I1010 16:20:53.947121 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.047460 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.047700 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfpcj\" (UniqueName: \"kubernetes.io/projected/398b7cc2-ff9d-4967-ad41-d89a87f831fe-kube-api-access-bfpcj\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.047722 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-scripts\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.047770 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/398b7cc2-ff9d-4967-ad41-d89a87f831fe-logs\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.047791 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.047842 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.047872 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/398b7cc2-ff9d-4967-ad41-d89a87f831fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.048027 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/398b7cc2-ff9d-4967-ad41-d89a87f831fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.048957 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/398b7cc2-ff9d-4967-ad41-d89a87f831fe-logs\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.054899 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.055590 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.057646 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.060835 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-scripts\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.069431 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfpcj\" (UniqueName: \"kubernetes.io/projected/398b7cc2-ff9d-4967-ad41-d89a87f831fe-kube-api-access-bfpcj\") pod \"cinder-api-0\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.123543 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.473362 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b757d79cc-8ltdz"] Oct 10 16:20:54 crc kubenswrapper[4788]: I1010 16:20:54.591060 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:20:55 crc kubenswrapper[4788]: I1010 16:20:55.234344 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:20:55 crc kubenswrapper[4788]: E1010 16:20:55.234830 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:20:55 crc kubenswrapper[4788]: I1010 16:20:55.312837 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"398b7cc2-ff9d-4967-ad41-d89a87f831fe","Type":"ContainerStarted","Data":"d394c63c71c9865571ec99f5a575460d5bb02efde29411e60b3790dc5cf1ba63"} Oct 10 16:20:55 crc kubenswrapper[4788]: I1010 16:20:55.312882 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"398b7cc2-ff9d-4967-ad41-d89a87f831fe","Type":"ContainerStarted","Data":"fcba09a3f540f74bc625dfce7ac5059f2a407a9ffcbe2ff8159aae81eb95e967"} Oct 10 16:20:55 crc kubenswrapper[4788]: I1010 16:20:55.315061 4788 generic.go:334] "Generic (PLEG): container finished" podID="e2664a06-b38b-49b9-8651-75098ac9d1f0" containerID="7fb9f350e115b134ad6783b35fe697699d2d135e30c15516b9bdf4f510aa953b" exitCode=0 Oct 10 16:20:55 crc kubenswrapper[4788]: I1010 16:20:55.315095 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" event={"ID":"e2664a06-b38b-49b9-8651-75098ac9d1f0","Type":"ContainerDied","Data":"7fb9f350e115b134ad6783b35fe697699d2d135e30c15516b9bdf4f510aa953b"} Oct 10 16:20:55 crc kubenswrapper[4788]: I1010 16:20:55.315113 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" event={"ID":"e2664a06-b38b-49b9-8651-75098ac9d1f0","Type":"ContainerStarted","Data":"53a88bd84d1bc7224169f11c60ef2913d00978526940abe206fbadd315782897"} Oct 10 16:20:56 crc kubenswrapper[4788]: I1010 16:20:56.325871 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"398b7cc2-ff9d-4967-ad41-d89a87f831fe","Type":"ContainerStarted","Data":"16f788c48be3e709c88b39e872796a44f86af03cb5831efe25d40dccbe76934c"} Oct 10 16:20:56 crc kubenswrapper[4788]: I1010 16:20:56.326248 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 10 16:20:56 crc kubenswrapper[4788]: I1010 16:20:56.328127 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" event={"ID":"e2664a06-b38b-49b9-8651-75098ac9d1f0","Type":"ContainerStarted","Data":"4d6b1a58dce6ea3ad07e4b51485fe47f33336351e659566710c66494646bff72"} Oct 10 16:20:56 crc kubenswrapper[4788]: I1010 16:20:56.328353 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:20:56 crc kubenswrapper[4788]: I1010 16:20:56.350673 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.3506240050000002 podStartE2EDuration="3.350624005s" podCreationTimestamp="2025-10-10 16:20:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:20:56.345802261 +0000 UTC m=+5758.795517829" watchObservedRunningTime="2025-10-10 16:20:56.350624005 +0000 UTC m=+5758.800339553" Oct 10 16:20:56 crc kubenswrapper[4788]: I1010 16:20:56.367620 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" podStartSLOduration=3.367603961 podStartE2EDuration="3.367603961s" podCreationTimestamp="2025-10-10 16:20:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:20:56.363909096 +0000 UTC m=+5758.813624644" watchObservedRunningTime="2025-10-10 16:20:56.367603961 +0000 UTC m=+5758.817319499" Oct 10 16:21:03 crc kubenswrapper[4788]: I1010 16:21:03.953463 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.061856 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6785b99c97-2j5bq"] Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.062686 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" podUID="ef80c08c-eef4-4e88-b18e-e43a38285f93" containerName="dnsmasq-dns" containerID="cri-o://18d44731daaac1d4886a45be352fdbe6636494e040147ea6e69da5b55759a4a4" gracePeriod=10 Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.431213 4788 generic.go:334] "Generic (PLEG): container finished" podID="ef80c08c-eef4-4e88-b18e-e43a38285f93" containerID="18d44731daaac1d4886a45be352fdbe6636494e040147ea6e69da5b55759a4a4" exitCode=0 Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.431677 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" event={"ID":"ef80c08c-eef4-4e88-b18e-e43a38285f93","Type":"ContainerDied","Data":"18d44731daaac1d4886a45be352fdbe6636494e040147ea6e69da5b55759a4a4"} Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.569269 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.655877 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-nb\") pod \"ef80c08c-eef4-4e88-b18e-e43a38285f93\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.656062 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-sb\") pod \"ef80c08c-eef4-4e88-b18e-e43a38285f93\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.656097 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82ddn\" (UniqueName: \"kubernetes.io/projected/ef80c08c-eef4-4e88-b18e-e43a38285f93-kube-api-access-82ddn\") pod \"ef80c08c-eef4-4e88-b18e-e43a38285f93\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.656122 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-dns-svc\") pod \"ef80c08c-eef4-4e88-b18e-e43a38285f93\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.656187 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-config\") pod \"ef80c08c-eef4-4e88-b18e-e43a38285f93\" (UID: \"ef80c08c-eef4-4e88-b18e-e43a38285f93\") " Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.678130 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef80c08c-eef4-4e88-b18e-e43a38285f93-kube-api-access-82ddn" (OuterVolumeSpecName: "kube-api-access-82ddn") pod "ef80c08c-eef4-4e88-b18e-e43a38285f93" (UID: "ef80c08c-eef4-4e88-b18e-e43a38285f93"). InnerVolumeSpecName "kube-api-access-82ddn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.718374 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef80c08c-eef4-4e88-b18e-e43a38285f93" (UID: "ef80c08c-eef4-4e88-b18e-e43a38285f93"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.719304 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef80c08c-eef4-4e88-b18e-e43a38285f93" (UID: "ef80c08c-eef4-4e88-b18e-e43a38285f93"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.730963 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef80c08c-eef4-4e88-b18e-e43a38285f93" (UID: "ef80c08c-eef4-4e88-b18e-e43a38285f93"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.736981 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-config" (OuterVolumeSpecName: "config") pod "ef80c08c-eef4-4e88-b18e-e43a38285f93" (UID: "ef80c08c-eef4-4e88-b18e-e43a38285f93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.759721 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.759755 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.760908 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.760918 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82ddn\" (UniqueName: \"kubernetes.io/projected/ef80c08c-eef4-4e88-b18e-e43a38285f93-kube-api-access-82ddn\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:04 crc kubenswrapper[4788]: I1010 16:21:04.760928 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef80c08c-eef4-4e88-b18e-e43a38285f93-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.449739 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" event={"ID":"ef80c08c-eef4-4e88-b18e-e43a38285f93","Type":"ContainerDied","Data":"b58ef9d21f2caacf8ea69a81d35e1ca43df6ecf17851eea14615871698ccec28"} Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.450207 4788 scope.go:117] "RemoveContainer" containerID="18d44731daaac1d4886a45be352fdbe6636494e040147ea6e69da5b55759a4a4" Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.450343 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6785b99c97-2j5bq" Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.480561 4788 scope.go:117] "RemoveContainer" containerID="ec7c7d6a469c2235e28569f122029dc36db8dc5fc4fd1b59cc882184cbd91206" Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.491719 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6785b99c97-2j5bq"] Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.496747 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6785b99c97-2j5bq"] Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.809270 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.809745 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-log" containerID="cri-o://3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb" gracePeriod=30 Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.809806 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-api" containerID="cri-o://5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2" gracePeriod=30 Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.819600 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.819992 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-log" containerID="cri-o://ff8ba4238c632907444b0341c93de733d6cd96167489db240b70f3285ac0b32f" gracePeriod=30 Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.820193 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-metadata" containerID="cri-o://960c4c7bf19c6b08d2b63577629db693d8befa34eead86a05bd574119e876057" gracePeriod=30 Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.830689 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.830962 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" containerName="nova-cell0-conductor-conductor" containerID="cri-o://3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04" gracePeriod=30 Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.844579 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.844958 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7b94d62f-25ad-4615-8d4e-863e5362b66f" containerName="nova-scheduler-scheduler" containerID="cri-o://b54ac40c052f0988b2e86ceb19ce6aab75391b5aa327c8fc731086fd2f3239ab" gracePeriod=30 Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.856972 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:21:05 crc kubenswrapper[4788]: I1010 16:21:05.857479 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0ae6b8020e8b0a2f6507160ffdffb85d8c06bde17f8669da74d72ea84027de67" gracePeriod=30 Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.150861 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.246749 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef80c08c-eef4-4e88-b18e-e43a38285f93" path="/var/lib/kubelet/pods/ef80c08c-eef4-4e88-b18e-e43a38285f93/volumes" Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.463370 4788 generic.go:334] "Generic (PLEG): container finished" podID="0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" containerID="0ae6b8020e8b0a2f6507160ffdffb85d8c06bde17f8669da74d72ea84027de67" exitCode=0 Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.463446 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e","Type":"ContainerDied","Data":"0ae6b8020e8b0a2f6507160ffdffb85d8c06bde17f8669da74d72ea84027de67"} Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.467471 4788 generic.go:334] "Generic (PLEG): container finished" podID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerID="ff8ba4238c632907444b0341c93de733d6cd96167489db240b70f3285ac0b32f" exitCode=143 Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.467546 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4","Type":"ContainerDied","Data":"ff8ba4238c632907444b0341c93de733d6cd96167489db240b70f3285ac0b32f"} Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.472404 4788 generic.go:334] "Generic (PLEG): container finished" podID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerID="3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb" exitCode=143 Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.472468 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4809b3c-2094-4a75-b601-792ebe980c9b","Type":"ContainerDied","Data":"3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb"} Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.650556 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.689970 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-config-data\") pod \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.690393 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-combined-ca-bundle\") pod \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.690560 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmvgq\" (UniqueName: \"kubernetes.io/projected/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-kube-api-access-hmvgq\") pod \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\" (UID: \"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e\") " Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.708389 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-kube-api-access-hmvgq" (OuterVolumeSpecName: "kube-api-access-hmvgq") pod "0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" (UID: "0a03b792-29c1-42fd-8e5a-3a6d04d21c3e"). InnerVolumeSpecName "kube-api-access-hmvgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.742361 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" (UID: "0a03b792-29c1-42fd-8e5a-3a6d04d21c3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.757400 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-config-data" (OuterVolumeSpecName: "config-data") pod "0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" (UID: "0a03b792-29c1-42fd-8e5a-3a6d04d21c3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.793109 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.793236 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:06 crc kubenswrapper[4788]: I1010 16:21:06.793251 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmvgq\" (UniqueName: \"kubernetes.io/projected/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e-kube-api-access-hmvgq\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.493585 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0a03b792-29c1-42fd-8e5a-3a6d04d21c3e","Type":"ContainerDied","Data":"ea15d58e56b69d6a65d4ea2fd670d21633ac37800ac100c28176e6a349971c1d"} Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.493641 4788 scope.go:117] "RemoveContainer" containerID="0ae6b8020e8b0a2f6507160ffdffb85d8c06bde17f8669da74d72ea84027de67" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.493776 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.519011 4788 generic.go:334] "Generic (PLEG): container finished" podID="7b94d62f-25ad-4615-8d4e-863e5362b66f" containerID="b54ac40c052f0988b2e86ceb19ce6aab75391b5aa327c8fc731086fd2f3239ab" exitCode=0 Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.519062 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7b94d62f-25ad-4615-8d4e-863e5362b66f","Type":"ContainerDied","Data":"b54ac40c052f0988b2e86ceb19ce6aab75391b5aa327c8fc731086fd2f3239ab"} Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.548383 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.553973 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.580634 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:21:07 crc kubenswrapper[4788]: E1010 16:21:07.581032 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef80c08c-eef4-4e88-b18e-e43a38285f93" containerName="dnsmasq-dns" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.581045 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef80c08c-eef4-4e88-b18e-e43a38285f93" containerName="dnsmasq-dns" Oct 10 16:21:07 crc kubenswrapper[4788]: E1010 16:21:07.581064 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef80c08c-eef4-4e88-b18e-e43a38285f93" containerName="init" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.581071 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef80c08c-eef4-4e88-b18e-e43a38285f93" containerName="init" Oct 10 16:21:07 crc kubenswrapper[4788]: E1010 16:21:07.581084 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.581092 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.581339 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" containerName="nova-cell1-novncproxy-novncproxy" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.581359 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef80c08c-eef4-4e88-b18e-e43a38285f93" containerName="dnsmasq-dns" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.582087 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.585367 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.592933 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.709090 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c017349-955a-4b6a-9ab3-b689d0316e6f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.709584 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c017349-955a-4b6a-9ab3-b689d0316e6f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.709634 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nrsz\" (UniqueName: \"kubernetes.io/projected/2c017349-955a-4b6a-9ab3-b689d0316e6f-kube-api-access-5nrsz\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.810873 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c017349-955a-4b6a-9ab3-b689d0316e6f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.810937 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nrsz\" (UniqueName: \"kubernetes.io/projected/2c017349-955a-4b6a-9ab3-b689d0316e6f-kube-api-access-5nrsz\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.810988 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c017349-955a-4b6a-9ab3-b689d0316e6f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.816328 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c017349-955a-4b6a-9ab3-b689d0316e6f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.816863 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c017349-955a-4b6a-9ab3-b689d0316e6f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.828247 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nrsz\" (UniqueName: \"kubernetes.io/projected/2c017349-955a-4b6a-9ab3-b689d0316e6f-kube-api-access-5nrsz\") pod \"nova-cell1-novncproxy-0\" (UID: \"2c017349-955a-4b6a-9ab3-b689d0316e6f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.881955 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.912476 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-config-data\") pod \"7b94d62f-25ad-4615-8d4e-863e5362b66f\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.912538 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58pnp\" (UniqueName: \"kubernetes.io/projected/7b94d62f-25ad-4615-8d4e-863e5362b66f-kube-api-access-58pnp\") pod \"7b94d62f-25ad-4615-8d4e-863e5362b66f\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.912801 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-combined-ca-bundle\") pod \"7b94d62f-25ad-4615-8d4e-863e5362b66f\" (UID: \"7b94d62f-25ad-4615-8d4e-863e5362b66f\") " Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.918051 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b94d62f-25ad-4615-8d4e-863e5362b66f-kube-api-access-58pnp" (OuterVolumeSpecName: "kube-api-access-58pnp") pod "7b94d62f-25ad-4615-8d4e-863e5362b66f" (UID: "7b94d62f-25ad-4615-8d4e-863e5362b66f"). InnerVolumeSpecName "kube-api-access-58pnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.918603 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.945463 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b94d62f-25ad-4615-8d4e-863e5362b66f" (UID: "7b94d62f-25ad-4615-8d4e-863e5362b66f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:07 crc kubenswrapper[4788]: I1010 16:21:07.948071 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-config-data" (OuterVolumeSpecName: "config-data") pod "7b94d62f-25ad-4615-8d4e-863e5362b66f" (UID: "7b94d62f-25ad-4615-8d4e-863e5362b66f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.015330 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.015363 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b94d62f-25ad-4615-8d4e-863e5362b66f-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.015373 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58pnp\" (UniqueName: \"kubernetes.io/projected/7b94d62f-25ad-4615-8d4e-863e5362b66f-kube-api-access-58pnp\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.245022 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a03b792-29c1-42fd-8e5a-3a6d04d21c3e" path="/var/lib/kubelet/pods/0a03b792-29c1-42fd-8e5a-3a6d04d21c3e/volumes" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.375236 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 10 16:21:08 crc kubenswrapper[4788]: W1010 16:21:08.387739 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c017349_955a_4b6a_9ab3_b689d0316e6f.slice/crio-b95575c203ee061edd4185f9733af531cc6b7409b65713a695c8475648932752 WatchSource:0}: Error finding container b95575c203ee061edd4185f9733af531cc6b7409b65713a695c8475648932752: Status 404 returned error can't find the container with id b95575c203ee061edd4185f9733af531cc6b7409b65713a695c8475648932752 Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.532254 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7b94d62f-25ad-4615-8d4e-863e5362b66f","Type":"ContainerDied","Data":"540b5be8f7ae77b694bf8f51ac6433dd21907318afca32c8b3db8c116ec2e9f5"} Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.532319 4788 scope.go:117] "RemoveContainer" containerID="b54ac40c052f0988b2e86ceb19ce6aab75391b5aa327c8fc731086fd2f3239ab" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.532440 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.534760 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2c017349-955a-4b6a-9ab3-b689d0316e6f","Type":"ContainerStarted","Data":"b95575c203ee061edd4185f9733af531cc6b7409b65713a695c8475648932752"} Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.574188 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.588236 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.611806 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:21:08 crc kubenswrapper[4788]: E1010 16:21:08.612367 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b94d62f-25ad-4615-8d4e-863e5362b66f" containerName="nova-scheduler-scheduler" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.612391 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b94d62f-25ad-4615-8d4e-863e5362b66f" containerName="nova-scheduler-scheduler" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.612793 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b94d62f-25ad-4615-8d4e-863e5362b66f" containerName="nova-scheduler-scheduler" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.613800 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.616278 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.626533 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.729609 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.732857 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-config-data\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.732980 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mbtp\" (UniqueName: \"kubernetes.io/projected/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-kube-api-access-8mbtp\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.835256 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-config-data\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.835367 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mbtp\" (UniqueName: \"kubernetes.io/projected/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-kube-api-access-8mbtp\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.835501 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.841814 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-config-data\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.851274 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.855886 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mbtp\" (UniqueName: \"kubernetes.io/projected/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-kube-api-access-8mbtp\") pod \"nova-scheduler-0\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.933984 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.960022 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": read tcp 10.217.0.2:33670->10.217.1.76:8775: read: connection reset by peer" Oct 10 16:21:08 crc kubenswrapper[4788]: I1010 16:21:08.960090 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": read tcp 10.217.0.2:33658->10.217.1.76:8775: read: connection reset by peer" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.031812 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.032359 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="bfe95d07-0b8f-49cf-9bb4-6817e193770e" containerName="nova-cell1-conductor-conductor" containerID="cri-o://44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251" gracePeriod=30 Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.196319 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.241034 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-config-data\") pod \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.241089 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf9zt\" (UniqueName: \"kubernetes.io/projected/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-kube-api-access-pf9zt\") pod \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.241125 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-combined-ca-bundle\") pod \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\" (UID: \"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.252305 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-kube-api-access-pf9zt" (OuterVolumeSpecName: "kube-api-access-pf9zt") pod "352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" (UID: "352d6b10-f034-4e21-9ed6-e1d5daa0c1d0"). InnerVolumeSpecName "kube-api-access-pf9zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.283909 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" (UID: "352d6b10-f034-4e21-9ed6-e1d5daa0c1d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.287750 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-config-data" (OuterVolumeSpecName: "config-data") pod "352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" (UID: "352d6b10-f034-4e21-9ed6-e1d5daa0c1d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.344968 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.345362 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf9zt\" (UniqueName: \"kubernetes.io/projected/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-kube-api-access-pf9zt\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.345377 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.539703 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.547527 4788 generic.go:334] "Generic (PLEG): container finished" podID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerID="960c4c7bf19c6b08d2b63577629db693d8befa34eead86a05bd574119e876057" exitCode=0 Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.547587 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4","Type":"ContainerDied","Data":"960c4c7bf19c6b08d2b63577629db693d8befa34eead86a05bd574119e876057"} Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.547614 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4","Type":"ContainerDied","Data":"cb075dea9e65dea2e1381a425f593ce2f9b29e0beedb3fdfcfac99363aa3e40b"} Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.547625 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb075dea9e65dea2e1381a425f593ce2f9b29e0beedb3fdfcfac99363aa3e40b" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.548909 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.549912 4788 generic.go:334] "Generic (PLEG): container finished" podID="352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" containerID="3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04" exitCode=0 Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.549958 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0","Type":"ContainerDied","Data":"3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04"} Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.549974 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"352d6b10-f034-4e21-9ed6-e1d5daa0c1d0","Type":"ContainerDied","Data":"e368c3ede74b655d93a69dfd0bca29af6f185d0a2d4872f06c5589872646fa78"} Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.549990 4788 scope.go:117] "RemoveContainer" containerID="3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.550108 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.553461 4788 generic.go:334] "Generic (PLEG): container finished" podID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerID="5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2" exitCode=0 Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.553503 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4809b3c-2094-4a75-b601-792ebe980c9b","Type":"ContainerDied","Data":"5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2"} Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.553521 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d4809b3c-2094-4a75-b601-792ebe980c9b","Type":"ContainerDied","Data":"3edb03b1bead459b65cd53fc64c2b5e8667c78c2706a278f082324ea93fe2185"} Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.553562 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.556385 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2c017349-955a-4b6a-9ab3-b689d0316e6f","Type":"ContainerStarted","Data":"7230c3567c20ed99aa13992cea3a994dd9dd72be3a7e94527675ccb03d49539e"} Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.601828 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.6017987849999997 podStartE2EDuration="2.601798785s" podCreationTimestamp="2025-10-10 16:21:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:09.582102639 +0000 UTC m=+5772.031818187" watchObservedRunningTime="2025-10-10 16:21:09.601798785 +0000 UTC m=+5772.051514333" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.606332 4788 scope.go:117] "RemoveContainer" containerID="3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04" Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.607246 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04\": container with ID starting with 3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04 not found: ID does not exist" containerID="3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.607288 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04"} err="failed to get container status \"3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04\": rpc error: code = NotFound desc = could not find container \"3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04\": container with ID starting with 3e05a517310a1e67c54ebd4ccce7ff849311fc990c5a2e6545973af76c327e04 not found: ID does not exist" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.607315 4788 scope.go:117] "RemoveContainer" containerID="5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.652466 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-config-data\") pod \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.652937 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.653026 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-logs\") pod \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.653097 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-combined-ca-bundle\") pod \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.653122 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbn6z\" (UniqueName: \"kubernetes.io/projected/d4809b3c-2094-4a75-b601-792ebe980c9b-kube-api-access-bbn6z\") pod \"d4809b3c-2094-4a75-b601-792ebe980c9b\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.653192 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-config-data\") pod \"d4809b3c-2094-4a75-b601-792ebe980c9b\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.653216 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxjzr\" (UniqueName: \"kubernetes.io/projected/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-kube-api-access-bxjzr\") pod \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\" (UID: \"ea232c34-bec6-4e90-a6fd-9a3ec7d520f4\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.653277 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4809b3c-2094-4a75-b601-792ebe980c9b-logs\") pod \"d4809b3c-2094-4a75-b601-792ebe980c9b\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.653314 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-combined-ca-bundle\") pod \"d4809b3c-2094-4a75-b601-792ebe980c9b\" (UID: \"d4809b3c-2094-4a75-b601-792ebe980c9b\") " Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.653551 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-logs" (OuterVolumeSpecName: "logs") pod "ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" (UID: "ea232c34-bec6-4e90-a6fd-9a3ec7d520f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.654041 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.654966 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4809b3c-2094-4a75-b601-792ebe980c9b-logs" (OuterVolumeSpecName: "logs") pod "d4809b3c-2094-4a75-b601-792ebe980c9b" (UID: "d4809b3c-2094-4a75-b601-792ebe980c9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.660857 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-kube-api-access-bxjzr" (OuterVolumeSpecName: "kube-api-access-bxjzr") pod "ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" (UID: "ea232c34-bec6-4e90-a6fd-9a3ec7d520f4"). InnerVolumeSpecName "kube-api-access-bxjzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.661821 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.666882 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4809b3c-2094-4a75-b601-792ebe980c9b-kube-api-access-bbn6z" (OuterVolumeSpecName: "kube-api-access-bbn6z") pod "d4809b3c-2094-4a75-b601-792ebe980c9b" (UID: "d4809b3c-2094-4a75-b601-792ebe980c9b"). InnerVolumeSpecName "kube-api-access-bbn6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.677754 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.677828 4788 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="bfe95d07-0b8f-49cf-9bb4-6817e193770e" containerName="nova-cell1-conductor-conductor" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.685324 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.694808 4788 scope.go:117] "RemoveContainer" containerID="3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.694845 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4809b3c-2094-4a75-b601-792ebe980c9b" (UID: "d4809b3c-2094-4a75-b601-792ebe980c9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.696461 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.697566 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" (UID: "ea232c34-bec6-4e90-a6fd-9a3ec7d520f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.698452 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-config-data" (OuterVolumeSpecName: "config-data") pod "d4809b3c-2094-4a75-b601-792ebe980c9b" (UID: "d4809b3c-2094-4a75-b601-792ebe980c9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.704392 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-config-data" (OuterVolumeSpecName: "config-data") pod "ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" (UID: "ea232c34-bec6-4e90-a6fd-9a3ec7d520f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.717617 4788 scope.go:117] "RemoveContainer" containerID="5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.717747 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.718975 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2\": container with ID starting with 5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2 not found: ID does not exist" containerID="5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.719020 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2"} err="failed to get container status \"5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2\": rpc error: code = NotFound desc = could not find container \"5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2\": container with ID starting with 5b7faa77f689b01c6d5d4e274d69d6594cfc022be4392e22e6468e7c0993a5e2 not found: ID does not exist" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.719047 4788 scope.go:117] "RemoveContainer" containerID="3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb" Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.721189 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb\": container with ID starting with 3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb not found: ID does not exist" containerID="3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.721238 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb"} err="failed to get container status \"3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb\": rpc error: code = NotFound desc = could not find container \"3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb\": container with ID starting with 3358ed2555d6d5788391f46c0a11e5335487605d586760b6ad5f6f1852639cbb not found: ID does not exist" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726019 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.726457 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-log" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726475 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-log" Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.726498 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-api" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726505 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-api" Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.726528 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-log" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726534 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-log" Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.726547 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-metadata" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726555 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-metadata" Oct 10 16:21:09 crc kubenswrapper[4788]: E1010 16:21:09.726570 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" containerName="nova-cell0-conductor-conductor" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726576 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" containerName="nova-cell0-conductor-conductor" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726731 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-metadata" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726761 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" containerName="nova-cell0-conductor-conductor" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726777 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-api" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726790 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" containerName="nova-api-log" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.726806 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" containerName="nova-metadata-log" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.727521 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.730425 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.740339 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.755469 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.755831 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9phv2\" (UniqueName: \"kubernetes.io/projected/f757adf6-3865-45a1-80ba-239090629e2b-kube-api-access-9phv2\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.756112 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.756354 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.756372 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.756387 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbn6z\" (UniqueName: \"kubernetes.io/projected/d4809b3c-2094-4a75-b601-792ebe980c9b-kube-api-access-bbn6z\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.756395 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.756404 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxjzr\" (UniqueName: \"kubernetes.io/projected/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4-kube-api-access-bxjzr\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.756412 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4809b3c-2094-4a75-b601-792ebe980c9b-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.756420 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4809b3c-2094-4a75-b601-792ebe980c9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.856958 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.857011 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.857041 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9phv2\" (UniqueName: \"kubernetes.io/projected/f757adf6-3865-45a1-80ba-239090629e2b-kube-api-access-9phv2\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.864485 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.864901 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.872883 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9phv2\" (UniqueName: \"kubernetes.io/projected/f757adf6-3865-45a1-80ba-239090629e2b-kube-api-access-9phv2\") pod \"nova-cell0-conductor-0\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.913636 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.935981 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.942236 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.943828 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.948911 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 16:21:09 crc kubenswrapper[4788]: I1010 16:21:09.950124 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.061783 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-config-data\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.061892 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwjcj\" (UniqueName: \"kubernetes.io/projected/fcc4fc46-31f4-45d1-9255-d3f5496b6840-kube-api-access-wwjcj\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.061946 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc4fc46-31f4-45d1-9255-d3f5496b6840-logs\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.062010 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.074596 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.164545 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-config-data\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.164618 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwjcj\" (UniqueName: \"kubernetes.io/projected/fcc4fc46-31f4-45d1-9255-d3f5496b6840-kube-api-access-wwjcj\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.164655 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc4fc46-31f4-45d1-9255-d3f5496b6840-logs\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.164702 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.166608 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc4fc46-31f4-45d1-9255-d3f5496b6840-logs\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.172890 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-config-data\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.178684 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.184597 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwjcj\" (UniqueName: \"kubernetes.io/projected/fcc4fc46-31f4-45d1-9255-d3f5496b6840-kube-api-access-wwjcj\") pod \"nova-api-0\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.235356 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:21:10 crc kubenswrapper[4788]: E1010 16:21:10.235567 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.258999 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352d6b10-f034-4e21-9ed6-e1d5daa0c1d0" path="/var/lib/kubelet/pods/352d6b10-f034-4e21-9ed6-e1d5daa0c1d0/volumes" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.260406 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b94d62f-25ad-4615-8d4e-863e5362b66f" path="/var/lib/kubelet/pods/7b94d62f-25ad-4615-8d4e-863e5362b66f/volumes" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.261060 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4809b3c-2094-4a75-b601-792ebe980c9b" path="/var/lib/kubelet/pods/d4809b3c-2094-4a75-b601-792ebe980c9b/volumes" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.262728 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.529603 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.584837 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f757adf6-3865-45a1-80ba-239090629e2b","Type":"ContainerStarted","Data":"afdf0c70421d58da3e20c29507b17570ffdc6d58a9cc0560eaa204fe596155e0"} Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.587406 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31","Type":"ContainerStarted","Data":"9f762d634176aaaa4210b783680d5baf3c7c6137f69bc900ce4efa3acb49dca4"} Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.587437 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31","Type":"ContainerStarted","Data":"26f454ba5fa2fa37f8eade1832c321e8798d905466e6da1fc974eeff277df928"} Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.591406 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.609761 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.609738588 podStartE2EDuration="2.609738588s" podCreationTimestamp="2025-10-10 16:21:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:10.601002574 +0000 UTC m=+5773.050718122" watchObservedRunningTime="2025-10-10 16:21:10.609738588 +0000 UTC m=+5773.059454126" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.720980 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.728713 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.744956 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.746922 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.749487 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.755313 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.840976 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 16:21:10 crc kubenswrapper[4788]: W1010 16:21:10.850006 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcc4fc46_31f4_45d1_9255_d3f5496b6840.slice/crio-dab4d8f4c0e015878d78c52ac537e48c9bdd37a6f2ee43e8df387637c10877d8 WatchSource:0}: Error finding container dab4d8f4c0e015878d78c52ac537e48c9bdd37a6f2ee43e8df387637c10877d8: Status 404 returned error can't find the container with id dab4d8f4c0e015878d78c52ac537e48c9bdd37a6f2ee43e8df387637c10877d8 Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.899792 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3178184c-97f1-4dfb-aea9-d969ac506208-logs\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.899914 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn8v2\" (UniqueName: \"kubernetes.io/projected/3178184c-97f1-4dfb-aea9-d969ac506208-kube-api-access-zn8v2\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.899940 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:10 crc kubenswrapper[4788]: I1010 16:21:10.900236 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-config-data\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.002329 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3178184c-97f1-4dfb-aea9-d969ac506208-logs\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.002407 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn8v2\" (UniqueName: \"kubernetes.io/projected/3178184c-97f1-4dfb-aea9-d969ac506208-kube-api-access-zn8v2\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.002426 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.002619 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-config-data\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.002715 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3178184c-97f1-4dfb-aea9-d969ac506208-logs\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.008333 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-config-data\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.009686 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.017897 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn8v2\" (UniqueName: \"kubernetes.io/projected/3178184c-97f1-4dfb-aea9-d969ac506208-kube-api-access-zn8v2\") pod \"nova-metadata-0\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.093331 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.556252 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.602589 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f757adf6-3865-45a1-80ba-239090629e2b","Type":"ContainerStarted","Data":"1cc7b83df429ca373bc48fe1d50a2e0e0360637ec494af9dda855dc6bda4e0f3"} Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.602681 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.604524 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3178184c-97f1-4dfb-aea9-d969ac506208","Type":"ContainerStarted","Data":"aa073bac6f334aa5eb006434de3790a057039394ca7bebe1184666c66a73e4c6"} Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.605791 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcc4fc46-31f4-45d1-9255-d3f5496b6840","Type":"ContainerStarted","Data":"4dda8c383b199b04672af35a66a42aa3033507d666c020e2dc7f9af35e09bda9"} Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.605817 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcc4fc46-31f4-45d1-9255-d3f5496b6840","Type":"ContainerStarted","Data":"a7d39af77ab3d3a330b37ba43345d98cd0ff2c271758721891e58d7630c041a1"} Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.605830 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcc4fc46-31f4-45d1-9255-d3f5496b6840","Type":"ContainerStarted","Data":"dab4d8f4c0e015878d78c52ac537e48c9bdd37a6f2ee43e8df387637c10877d8"} Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.623345 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.623327807 podStartE2EDuration="2.623327807s" podCreationTimestamp="2025-10-10 16:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:11.616316467 +0000 UTC m=+5774.066032015" watchObservedRunningTime="2025-10-10 16:21:11.623327807 +0000 UTC m=+5774.073043355" Oct 10 16:21:11 crc kubenswrapper[4788]: I1010 16:21:11.636527 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.636508546 podStartE2EDuration="2.636508546s" podCreationTimestamp="2025-10-10 16:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:11.635182681 +0000 UTC m=+5774.084898290" watchObservedRunningTime="2025-10-10 16:21:11.636508546 +0000 UTC m=+5774.086224094" Oct 10 16:21:12 crc kubenswrapper[4788]: I1010 16:21:12.245379 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea232c34-bec6-4e90-a6fd-9a3ec7d520f4" path="/var/lib/kubelet/pods/ea232c34-bec6-4e90-a6fd-9a3ec7d520f4/volumes" Oct 10 16:21:12 crc kubenswrapper[4788]: I1010 16:21:12.616496 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3178184c-97f1-4dfb-aea9-d969ac506208","Type":"ContainerStarted","Data":"b1f6730c0bfab5eb494145f52bca89a2c0059f624a0cfe64c7b1cf4ee89b4b39"} Oct 10 16:21:12 crc kubenswrapper[4788]: I1010 16:21:12.616548 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3178184c-97f1-4dfb-aea9-d969ac506208","Type":"ContainerStarted","Data":"f6bbec90a1956f4b6b68625c173efca3d4de9f6019ffb56152a6cd2b1cd11cba"} Oct 10 16:21:12 crc kubenswrapper[4788]: I1010 16:21:12.646379 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.646355989 podStartE2EDuration="2.646355989s" podCreationTimestamp="2025-10-10 16:21:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:12.642356446 +0000 UTC m=+5775.092072034" watchObservedRunningTime="2025-10-10 16:21:12.646355989 +0000 UTC m=+5775.096071537" Oct 10 16:21:12 crc kubenswrapper[4788]: I1010 16:21:12.919037 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:13 crc kubenswrapper[4788]: I1010 16:21:13.934933 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.294758 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.477929 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-config-data\") pod \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.477999 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-combined-ca-bundle\") pod \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.478264 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl7z8\" (UniqueName: \"kubernetes.io/projected/bfe95d07-0b8f-49cf-9bb4-6817e193770e-kube-api-access-gl7z8\") pod \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\" (UID: \"bfe95d07-0b8f-49cf-9bb4-6817e193770e\") " Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.489342 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe95d07-0b8f-49cf-9bb4-6817e193770e-kube-api-access-gl7z8" (OuterVolumeSpecName: "kube-api-access-gl7z8") pod "bfe95d07-0b8f-49cf-9bb4-6817e193770e" (UID: "bfe95d07-0b8f-49cf-9bb4-6817e193770e"). InnerVolumeSpecName "kube-api-access-gl7z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.508076 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-config-data" (OuterVolumeSpecName: "config-data") pod "bfe95d07-0b8f-49cf-9bb4-6817e193770e" (UID: "bfe95d07-0b8f-49cf-9bb4-6817e193770e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.508628 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfe95d07-0b8f-49cf-9bb4-6817e193770e" (UID: "bfe95d07-0b8f-49cf-9bb4-6817e193770e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.581902 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.582437 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe95d07-0b8f-49cf-9bb4-6817e193770e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.582571 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl7z8\" (UniqueName: \"kubernetes.io/projected/bfe95d07-0b8f-49cf-9bb4-6817e193770e-kube-api-access-gl7z8\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.641247 4788 generic.go:334] "Generic (PLEG): container finished" podID="bfe95d07-0b8f-49cf-9bb4-6817e193770e" containerID="44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251" exitCode=0 Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.641269 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.641300 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bfe95d07-0b8f-49cf-9bb4-6817e193770e","Type":"ContainerDied","Data":"44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251"} Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.642907 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bfe95d07-0b8f-49cf-9bb4-6817e193770e","Type":"ContainerDied","Data":"60b9dc4c95e75c0e30917f0275b33cbecc0bf12a9c3ccb49377baa380f4c4179"} Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.642942 4788 scope.go:117] "RemoveContainer" containerID="44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.674222 4788 scope.go:117] "RemoveContainer" containerID="44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251" Oct 10 16:21:14 crc kubenswrapper[4788]: E1010 16:21:14.675642 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251\": container with ID starting with 44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251 not found: ID does not exist" containerID="44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.675693 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251"} err="failed to get container status \"44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251\": rpc error: code = NotFound desc = could not find container \"44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251\": container with ID starting with 44bef3862cb1531c0174d408f74e510e0cccc7b3b0ca7b1b56ae0e56ee4c1251 not found: ID does not exist" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.737655 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.753843 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.763091 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:21:14 crc kubenswrapper[4788]: E1010 16:21:14.763751 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe95d07-0b8f-49cf-9bb4-6817e193770e" containerName="nova-cell1-conductor-conductor" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.763782 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe95d07-0b8f-49cf-9bb4-6817e193770e" containerName="nova-cell1-conductor-conductor" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.764086 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe95d07-0b8f-49cf-9bb4-6817e193770e" containerName="nova-cell1-conductor-conductor" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.765183 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.769434 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.775266 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.894983 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.895690 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmwtl\" (UniqueName: \"kubernetes.io/projected/79194ba7-bc68-4165-8ab1-a6956df339cd-kube-api-access-fmwtl\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.895817 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:14 crc kubenswrapper[4788]: I1010 16:21:14.998748 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:14.998986 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmwtl\" (UniqueName: \"kubernetes.io/projected/79194ba7-bc68-4165-8ab1-a6956df339cd-kube-api-access-fmwtl\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:14.999212 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:15.004674 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:15.005679 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:15.019583 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmwtl\" (UniqueName: \"kubernetes.io/projected/79194ba7-bc68-4165-8ab1-a6956df339cd-kube-api-access-fmwtl\") pod \"nova-cell1-conductor-0\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:15.093847 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:15.122402 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:15.635823 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 16:21:15 crc kubenswrapper[4788]: W1010 16:21:15.640393 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79194ba7_bc68_4165_8ab1_a6956df339cd.slice/crio-409417680e7d64ef1b187210be125e9ea4d4b05a8ed19babc3db458fcbb4bd19 WatchSource:0}: Error finding container 409417680e7d64ef1b187210be125e9ea4d4b05a8ed19babc3db458fcbb4bd19: Status 404 returned error can't find the container with id 409417680e7d64ef1b187210be125e9ea4d4b05a8ed19babc3db458fcbb4bd19 Oct 10 16:21:15 crc kubenswrapper[4788]: I1010 16:21:15.655700 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"79194ba7-bc68-4165-8ab1-a6956df339cd","Type":"ContainerStarted","Data":"409417680e7d64ef1b187210be125e9ea4d4b05a8ed19babc3db458fcbb4bd19"} Oct 10 16:21:16 crc kubenswrapper[4788]: I1010 16:21:16.094490 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 16:21:16 crc kubenswrapper[4788]: I1010 16:21:16.094888 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 16:21:16 crc kubenswrapper[4788]: I1010 16:21:16.245645 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe95d07-0b8f-49cf-9bb4-6817e193770e" path="/var/lib/kubelet/pods/bfe95d07-0b8f-49cf-9bb4-6817e193770e/volumes" Oct 10 16:21:16 crc kubenswrapper[4788]: I1010 16:21:16.669379 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"79194ba7-bc68-4165-8ab1-a6956df339cd","Type":"ContainerStarted","Data":"0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b"} Oct 10 16:21:16 crc kubenswrapper[4788]: I1010 16:21:16.669564 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:16 crc kubenswrapper[4788]: I1010 16:21:16.698101 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.698071296 podStartE2EDuration="2.698071296s" podCreationTimestamp="2025-10-10 16:21:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:16.689354992 +0000 UTC m=+5779.139070540" watchObservedRunningTime="2025-10-10 16:21:16.698071296 +0000 UTC m=+5779.147786854" Oct 10 16:21:17 crc kubenswrapper[4788]: I1010 16:21:17.919933 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:17 crc kubenswrapper[4788]: I1010 16:21:17.941515 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:18 crc kubenswrapper[4788]: I1010 16:21:18.707127 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 10 16:21:18 crc kubenswrapper[4788]: I1010 16:21:18.934899 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 16:21:18 crc kubenswrapper[4788]: I1010 16:21:18.967242 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 16:21:19 crc kubenswrapper[4788]: I1010 16:21:19.740525 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 16:21:20 crc kubenswrapper[4788]: I1010 16:21:20.136440 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 10 16:21:20 crc kubenswrapper[4788]: I1010 16:21:20.263564 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 16:21:20 crc kubenswrapper[4788]: I1010 16:21:20.264165 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 16:21:21 crc kubenswrapper[4788]: I1010 16:21:21.093946 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 16:21:21 crc kubenswrapper[4788]: I1010 16:21:21.093990 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 16:21:21 crc kubenswrapper[4788]: I1010 16:21:21.233684 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:21:21 crc kubenswrapper[4788]: E1010 16:21:21.233933 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:21:21 crc kubenswrapper[4788]: I1010 16:21:21.346317 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.87:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:21:21 crc kubenswrapper[4788]: I1010 16:21:21.346356 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.87:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:21:22 crc kubenswrapper[4788]: I1010 16:21:22.176331 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.88:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:21:22 crc kubenswrapper[4788]: I1010 16:21:22.176378 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.88:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 16:21:25 crc kubenswrapper[4788]: I1010 16:21:25.997994 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.001966 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.015522 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.023219 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.041832 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5bkd\" (UniqueName: \"kubernetes.io/projected/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-kube-api-access-d5bkd\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.041907 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.041936 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.041974 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.042053 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.042072 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.144192 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5bkd\" (UniqueName: \"kubernetes.io/projected/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-kube-api-access-d5bkd\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.145814 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.145858 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.145907 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.146023 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.146045 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.146027 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.152978 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.153272 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.153569 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.155902 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.162020 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5bkd\" (UniqueName: \"kubernetes.io/projected/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-kube-api-access-d5bkd\") pod \"cinder-scheduler-0\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.349976 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 16:21:26 crc kubenswrapper[4788]: I1010 16:21:26.830769 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:27 crc kubenswrapper[4788]: I1010 16:21:27.567024 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:21:27 crc kubenswrapper[4788]: I1010 16:21:27.567896 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api-log" containerID="cri-o://d394c63c71c9865571ec99f5a575460d5bb02efde29411e60b3790dc5cf1ba63" gracePeriod=30 Oct 10 16:21:27 crc kubenswrapper[4788]: I1010 16:21:27.568045 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api" containerID="cri-o://16f788c48be3e709c88b39e872796a44f86af03cb5831efe25d40dccbe76934c" gracePeriod=30 Oct 10 16:21:27 crc kubenswrapper[4788]: I1010 16:21:27.854492 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd","Type":"ContainerStarted","Data":"b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83"} Oct 10 16:21:27 crc kubenswrapper[4788]: I1010 16:21:27.854555 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd","Type":"ContainerStarted","Data":"83028397f6bdb6dc024faa3692c66a5ba971f0491894e30abbcef26d3fa24534"} Oct 10 16:21:27 crc kubenswrapper[4788]: I1010 16:21:27.860432 4788 generic.go:334] "Generic (PLEG): container finished" podID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerID="d394c63c71c9865571ec99f5a575460d5bb02efde29411e60b3790dc5cf1ba63" exitCode=143 Oct 10 16:21:27 crc kubenswrapper[4788]: I1010 16:21:27.860477 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"398b7cc2-ff9d-4967-ad41-d89a87f831fe","Type":"ContainerDied","Data":"d394c63c71c9865571ec99f5a575460d5bb02efde29411e60b3790dc5cf1ba63"} Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.063890 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.066105 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.069686 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.080836 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192264 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192335 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192452 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192514 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192549 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192567 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192589 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192688 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192805 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/53564aa8-cbc4-4373-8330-ab61e819e9b7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192848 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192892 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192908 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192923 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.192942 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-run\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.193079 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.193131 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-956pz\" (UniqueName: \"kubernetes.io/projected/53564aa8-cbc4-4373-8330-ab61e819e9b7-kube-api-access-956pz\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295100 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295163 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295184 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295205 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295234 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295276 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/53564aa8-cbc4-4373-8330-ab61e819e9b7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295294 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295330 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295343 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295357 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295376 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-run\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295392 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295422 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-956pz\" (UniqueName: \"kubernetes.io/projected/53564aa8-cbc4-4373-8330-ab61e819e9b7-kube-api-access-956pz\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295456 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295485 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295518 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.295784 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296380 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296402 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296470 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296496 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296507 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296572 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296539 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-run\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296527 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.296670 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/53564aa8-cbc4-4373-8330-ab61e819e9b7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.301798 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.304519 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.309188 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.323045 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53564aa8-cbc4-4373-8330-ab61e819e9b7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.323832 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/53564aa8-cbc4-4373-8330-ab61e819e9b7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.326781 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-956pz\" (UniqueName: \"kubernetes.io/projected/53564aa8-cbc4-4373-8330-ab61e819e9b7-kube-api-access-956pz\") pod \"cinder-volume-volume1-0\" (UID: \"53564aa8-cbc4-4373-8330-ab61e819e9b7\") " pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.391912 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.725389 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.727363 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.732298 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.743957 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.805713 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-config-data\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.805794 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-run\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.805818 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7st6\" (UniqueName: \"kubernetes.io/projected/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-kube-api-access-v7st6\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.805876 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.805906 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-dev\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.805949 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.805981 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806026 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806049 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-lib-modules\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806109 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-scripts\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806164 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806193 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806208 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806260 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806281 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-sys\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.806298 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-ceph\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.874192 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd","Type":"ContainerStarted","Data":"d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b"} Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.898934 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.898913903 podStartE2EDuration="3.898913903s" podCreationTimestamp="2025-10-10 16:21:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:28.892673413 +0000 UTC m=+5791.342388961" watchObservedRunningTime="2025-10-10 16:21:28.898913903 +0000 UTC m=+5791.348629451" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.910859 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-config-data\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.910958 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-run\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.910994 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7st6\" (UniqueName: \"kubernetes.io/projected/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-kube-api-access-v7st6\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911046 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911078 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-run\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911092 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-dev\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911176 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-dev\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911247 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911318 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911351 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911385 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-lib-modules\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911439 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-scripts\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911480 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911494 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911550 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911563 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911580 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911604 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-lib-modules\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911673 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911734 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-sys\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.911759 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-ceph\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.912067 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.912091 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.912163 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-sys\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.912192 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.918504 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-scripts\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.921818 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-ceph\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.922725 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-config-data\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.924245 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.925410 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.932242 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7st6\" (UniqueName: \"kubernetes.io/projected/f6ef5c29-c7b8-4181-ab0d-4725a545a8a7-kube-api-access-v7st6\") pod \"cinder-backup-0\" (UID: \"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7\") " pod="openstack/cinder-backup-0" Oct 10 16:21:28 crc kubenswrapper[4788]: I1010 16:21:28.989871 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 10 16:21:28 crc kubenswrapper[4788]: W1010 16:21:28.996373 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53564aa8_cbc4_4373_8330_ab61e819e9b7.slice/crio-39643dc262b3bd24ce3bc1f35aa56e32fc0a577db0259bdcfb14bbeb3e568a97 WatchSource:0}: Error finding container 39643dc262b3bd24ce3bc1f35aa56e32fc0a577db0259bdcfb14bbeb3e568a97: Status 404 returned error can't find the container with id 39643dc262b3bd24ce3bc1f35aa56e32fc0a577db0259bdcfb14bbeb3e568a97 Oct 10 16:21:29 crc kubenswrapper[4788]: I1010 16:21:29.000806 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:21:29 crc kubenswrapper[4788]: I1010 16:21:29.058240 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 10 16:21:29 crc kubenswrapper[4788]: I1010 16:21:29.680900 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 10 16:21:29 crc kubenswrapper[4788]: I1010 16:21:29.890507 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7","Type":"ContainerStarted","Data":"e006901ec0b7c337345230589a2d60e46bf361004bff57065158de7313839c44"} Oct 10 16:21:29 crc kubenswrapper[4788]: I1010 16:21:29.892540 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"53564aa8-cbc4-4373-8330-ab61e819e9b7","Type":"ContainerStarted","Data":"39643dc262b3bd24ce3bc1f35aa56e32fc0a577db0259bdcfb14bbeb3e568a97"} Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.267421 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.268802 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.269998 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.279059 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.722735 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.83:8776/healthcheck\": read tcp 10.217.0.2:58878->10.217.1.83:8776: read: connection reset by peer" Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.907874 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"53564aa8-cbc4-4373-8330-ab61e819e9b7","Type":"ContainerStarted","Data":"9ffa640d6af9095cde5234f82d9dbbbd2bfcf823b6060765959f54be71e7d240"} Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.907930 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"53564aa8-cbc4-4373-8330-ab61e819e9b7","Type":"ContainerStarted","Data":"77af5b1b6b63241b76b329356f6ac17ed84950ce39939ee73449ca4b009cbdda"} Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.910999 4788 generic.go:334] "Generic (PLEG): container finished" podID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerID="16f788c48be3e709c88b39e872796a44f86af03cb5831efe25d40dccbe76934c" exitCode=0 Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.911218 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"398b7cc2-ff9d-4967-ad41-d89a87f831fe","Type":"ContainerDied","Data":"16f788c48be3e709c88b39e872796a44f86af03cb5831efe25d40dccbe76934c"} Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.911429 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.922511 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 16:21:30 crc kubenswrapper[4788]: I1010 16:21:30.950519 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=1.901246413 podStartE2EDuration="2.950496098s" podCreationTimestamp="2025-10-10 16:21:28 +0000 UTC" firstStartedPulling="2025-10-10 16:21:29.000425511 +0000 UTC m=+5791.450141079" lastFinishedPulling="2025-10-10 16:21:30.049675216 +0000 UTC m=+5792.499390764" observedRunningTime="2025-10-10 16:21:30.947043359 +0000 UTC m=+5793.396758917" watchObservedRunningTime="2025-10-10 16:21:30.950496098 +0000 UTC m=+5793.400211646" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.099613 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.100188 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.107379 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.114219 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.201043 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.273607 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/398b7cc2-ff9d-4967-ad41-d89a87f831fe-etc-machine-id\") pod \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.273680 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-combined-ca-bundle\") pod \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.273731 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data\") pod \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.273735 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/398b7cc2-ff9d-4967-ad41-d89a87f831fe-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "398b7cc2-ff9d-4967-ad41-d89a87f831fe" (UID: "398b7cc2-ff9d-4967-ad41-d89a87f831fe"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.273817 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-scripts\") pod \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.273886 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfpcj\" (UniqueName: \"kubernetes.io/projected/398b7cc2-ff9d-4967-ad41-d89a87f831fe-kube-api-access-bfpcj\") pod \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.273943 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data-custom\") pod \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.274045 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/398b7cc2-ff9d-4967-ad41-d89a87f831fe-logs\") pod \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\" (UID: \"398b7cc2-ff9d-4967-ad41-d89a87f831fe\") " Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.274660 4788 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/398b7cc2-ff9d-4967-ad41-d89a87f831fe-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.275433 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/398b7cc2-ff9d-4967-ad41-d89a87f831fe-logs" (OuterVolumeSpecName: "logs") pod "398b7cc2-ff9d-4967-ad41-d89a87f831fe" (UID: "398b7cc2-ff9d-4967-ad41-d89a87f831fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.283890 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "398b7cc2-ff9d-4967-ad41-d89a87f831fe" (UID: "398b7cc2-ff9d-4967-ad41-d89a87f831fe"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.284394 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-scripts" (OuterVolumeSpecName: "scripts") pod "398b7cc2-ff9d-4967-ad41-d89a87f831fe" (UID: "398b7cc2-ff9d-4967-ad41-d89a87f831fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.291795 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/398b7cc2-ff9d-4967-ad41-d89a87f831fe-kube-api-access-bfpcj" (OuterVolumeSpecName: "kube-api-access-bfpcj") pod "398b7cc2-ff9d-4967-ad41-d89a87f831fe" (UID: "398b7cc2-ff9d-4967-ad41-d89a87f831fe"). InnerVolumeSpecName "kube-api-access-bfpcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.317461 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "398b7cc2-ff9d-4967-ad41-d89a87f831fe" (UID: "398b7cc2-ff9d-4967-ad41-d89a87f831fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.331462 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data" (OuterVolumeSpecName: "config-data") pod "398b7cc2-ff9d-4967-ad41-d89a87f831fe" (UID: "398b7cc2-ff9d-4967-ad41-d89a87f831fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.350951 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.376932 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.377107 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfpcj\" (UniqueName: \"kubernetes.io/projected/398b7cc2-ff9d-4967-ad41-d89a87f831fe-kube-api-access-bfpcj\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.377122 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.377151 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/398b7cc2-ff9d-4967-ad41-d89a87f831fe-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.377159 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.377168 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398b7cc2-ff9d-4967-ad41-d89a87f831fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.935725 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7","Type":"ContainerStarted","Data":"e90796ae2817e1ea091d5be32e86be755d184987d804deaf6d4a5fee3897e865"} Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.936252 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"f6ef5c29-c7b8-4181-ab0d-4725a545a8a7","Type":"ContainerStarted","Data":"80e8ba565b67ee3149c03deb2aa9f0a1c01107d0bab384149969f28b4b1137bd"} Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.939693 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"398b7cc2-ff9d-4967-ad41-d89a87f831fe","Type":"ContainerDied","Data":"fcba09a3f540f74bc625dfce7ac5059f2a407a9ffcbe2ff8159aae81eb95e967"} Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.939772 4788 scope.go:117] "RemoveContainer" containerID="16f788c48be3e709c88b39e872796a44f86af03cb5831efe25d40dccbe76934c" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.940262 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.969067 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.926993234 podStartE2EDuration="3.969039544s" podCreationTimestamp="2025-10-10 16:21:28 +0000 UTC" firstStartedPulling="2025-10-10 16:21:29.677938266 +0000 UTC m=+5792.127653814" lastFinishedPulling="2025-10-10 16:21:30.719984566 +0000 UTC m=+5793.169700124" observedRunningTime="2025-10-10 16:21:31.963221784 +0000 UTC m=+5794.412937352" watchObservedRunningTime="2025-10-10 16:21:31.969039544 +0000 UTC m=+5794.418755092" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.985935 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.990550 4788 scope.go:117] "RemoveContainer" containerID="d394c63c71c9865571ec99f5a575460d5bb02efde29411e60b3790dc5cf1ba63" Oct 10 16:21:31 crc kubenswrapper[4788]: I1010 16:21:31.995173 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.017218 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:21:32 crc kubenswrapper[4788]: E1010 16:21:32.018012 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api-log" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.018033 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api-log" Oct 10 16:21:32 crc kubenswrapper[4788]: E1010 16:21:32.018063 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.018076 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.019219 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.019266 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" containerName="cinder-api-log" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.076228 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.080986 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.116284 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.117228 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-config-data-custom\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.117370 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15109044-050e-4ac9-b05e-8d5b80bf2f69-etc-machine-id\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.117422 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.117506 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15109044-050e-4ac9-b05e-8d5b80bf2f69-logs\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.117675 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-scripts\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.117728 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn4t4\" (UniqueName: \"kubernetes.io/projected/15109044-050e-4ac9-b05e-8d5b80bf2f69-kube-api-access-pn4t4\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.118017 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-config-data\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.224876 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15109044-050e-4ac9-b05e-8d5b80bf2f69-logs\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.225361 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-scripts\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.225405 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn4t4\" (UniqueName: \"kubernetes.io/projected/15109044-050e-4ac9-b05e-8d5b80bf2f69-kube-api-access-pn4t4\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.225448 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-config-data\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.225518 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-config-data-custom\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.225610 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15109044-050e-4ac9-b05e-8d5b80bf2f69-etc-machine-id\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.225660 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.227681 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15109044-050e-4ac9-b05e-8d5b80bf2f69-logs\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.229039 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15109044-050e-4ac9-b05e-8d5b80bf2f69-etc-machine-id\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.237292 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.238564 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-scripts\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.242132 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-config-data\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.244742 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15109044-050e-4ac9-b05e-8d5b80bf2f69-config-data-custom\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.248452 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn4t4\" (UniqueName: \"kubernetes.io/projected/15109044-050e-4ac9-b05e-8d5b80bf2f69-kube-api-access-pn4t4\") pod \"cinder-api-0\" (UID: \"15109044-050e-4ac9-b05e-8d5b80bf2f69\") " pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.252048 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="398b7cc2-ff9d-4967-ad41-d89a87f831fe" path="/var/lib/kubelet/pods/398b7cc2-ff9d-4967-ad41-d89a87f831fe/volumes" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.419221 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.919469 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 10 16:21:32 crc kubenswrapper[4788]: W1010 16:21:32.933066 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15109044_050e_4ac9_b05e_8d5b80bf2f69.slice/crio-8f2066fd58bb9336f233210eb02e6f2dd33b43439bd8003d73e5009de77dca70 WatchSource:0}: Error finding container 8f2066fd58bb9336f233210eb02e6f2dd33b43439bd8003d73e5009de77dca70: Status 404 returned error can't find the container with id 8f2066fd58bb9336f233210eb02e6f2dd33b43439bd8003d73e5009de77dca70 Oct 10 16:21:32 crc kubenswrapper[4788]: I1010 16:21:32.959868 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"15109044-050e-4ac9-b05e-8d5b80bf2f69","Type":"ContainerStarted","Data":"8f2066fd58bb9336f233210eb02e6f2dd33b43439bd8003d73e5009de77dca70"} Oct 10 16:21:33 crc kubenswrapper[4788]: I1010 16:21:33.393086 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:33 crc kubenswrapper[4788]: I1010 16:21:33.990635 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"15109044-050e-4ac9-b05e-8d5b80bf2f69","Type":"ContainerStarted","Data":"aad32c4e1a491151b33acdbe09d83b2c5f1a19e697f4cf8746d75b7166f0a9f7"} Oct 10 16:21:34 crc kubenswrapper[4788]: I1010 16:21:34.059177 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 10 16:21:35 crc kubenswrapper[4788]: I1010 16:21:35.005635 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"15109044-050e-4ac9-b05e-8d5b80bf2f69","Type":"ContainerStarted","Data":"a632615b483ea76b6d9259caba1c420469a753f43f6a0a9f4045510cda0df883"} Oct 10 16:21:35 crc kubenswrapper[4788]: I1010 16:21:35.006271 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 10 16:21:35 crc kubenswrapper[4788]: I1010 16:21:35.049971 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.049929162 podStartE2EDuration="4.049929162s" podCreationTimestamp="2025-10-10 16:21:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:35.036598409 +0000 UTC m=+5797.486313967" watchObservedRunningTime="2025-10-10 16:21:35.049929162 +0000 UTC m=+5797.499644730" Oct 10 16:21:35 crc kubenswrapper[4788]: I1010 16:21:35.234282 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:21:35 crc kubenswrapper[4788]: E1010 16:21:35.234600 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:21:36 crc kubenswrapper[4788]: I1010 16:21:36.555107 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 10 16:21:36 crc kubenswrapper[4788]: I1010 16:21:36.633278 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:37 crc kubenswrapper[4788]: I1010 16:21:37.027239 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerName="cinder-scheduler" containerID="cri-o://b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83" gracePeriod=30 Oct 10 16:21:37 crc kubenswrapper[4788]: I1010 16:21:37.028112 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerName="probe" containerID="cri-o://d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b" gracePeriod=30 Oct 10 16:21:38 crc kubenswrapper[4788]: I1010 16:21:38.044895 4788 generic.go:334] "Generic (PLEG): container finished" podID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerID="d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b" exitCode=0 Oct 10 16:21:38 crc kubenswrapper[4788]: I1010 16:21:38.044980 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd","Type":"ContainerDied","Data":"d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b"} Oct 10 16:21:38 crc kubenswrapper[4788]: I1010 16:21:38.621180 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.304030 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.688215 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.833677 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5bkd\" (UniqueName: \"kubernetes.io/projected/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-kube-api-access-d5bkd\") pod \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.833735 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data-custom\") pod \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.833801 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-etc-machine-id\") pod \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.833823 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data\") pod \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.833864 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-scripts\") pod \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.833888 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-combined-ca-bundle\") pod \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\" (UID: \"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd\") " Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.834048 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" (UID: "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.834321 4788 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.840895 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-kube-api-access-d5bkd" (OuterVolumeSpecName: "kube-api-access-d5bkd") pod "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" (UID: "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd"). InnerVolumeSpecName "kube-api-access-d5bkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.842513 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-scripts" (OuterVolumeSpecName: "scripts") pod "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" (UID: "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.845448 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" (UID: "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.893863 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" (UID: "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.937080 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5bkd\" (UniqueName: \"kubernetes.io/projected/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-kube-api-access-d5bkd\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.937126 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.937157 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.937171 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:39 crc kubenswrapper[4788]: I1010 16:21:39.950654 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data" (OuterVolumeSpecName: "config-data") pod "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" (UID: "1a5051a2-fcf0-47f3-add1-e1e5a0d226bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.039730 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.092360 4788 generic.go:334] "Generic (PLEG): container finished" podID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerID="b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83" exitCode=0 Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.092423 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd","Type":"ContainerDied","Data":"b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83"} Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.092472 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a5051a2-fcf0-47f3-add1-e1e5a0d226bd","Type":"ContainerDied","Data":"83028397f6bdb6dc024faa3692c66a5ba971f0491894e30abbcef26d3fa24534"} Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.092493 4788 scope.go:117] "RemoveContainer" containerID="d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.092786 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.121799 4788 scope.go:117] "RemoveContainer" containerID="b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.153905 4788 scope.go:117] "RemoveContainer" containerID="d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b" Oct 10 16:21:40 crc kubenswrapper[4788]: E1010 16:21:40.154613 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b\": container with ID starting with d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b not found: ID does not exist" containerID="d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.154729 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b"} err="failed to get container status \"d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b\": rpc error: code = NotFound desc = could not find container \"d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b\": container with ID starting with d518e004f24fcf86a065ff5a30323103339119bfb989e46920dae2a1ad89501b not found: ID does not exist" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.154791 4788 scope.go:117] "RemoveContainer" containerID="b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83" Oct 10 16:21:40 crc kubenswrapper[4788]: E1010 16:21:40.155175 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83\": container with ID starting with b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83 not found: ID does not exist" containerID="b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.155212 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83"} err="failed to get container status \"b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83\": rpc error: code = NotFound desc = could not find container \"b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83\": container with ID starting with b9c2c9ef2f1b2b69ad1b81798a70d5b6dc23adc0bb5faf4bfb28d4110fae7b83 not found: ID does not exist" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.160495 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.176846 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.186076 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:40 crc kubenswrapper[4788]: E1010 16:21:40.186667 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerName="cinder-scheduler" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.186699 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerName="cinder-scheduler" Oct 10 16:21:40 crc kubenswrapper[4788]: E1010 16:21:40.186719 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerName="probe" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.186728 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerName="probe" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.186986 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerName="cinder-scheduler" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.187006 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" containerName="probe" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.188394 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.192709 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.194338 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.256935 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a5051a2-fcf0-47f3-add1-e1e5a0d226bd" path="/var/lib/kubelet/pods/1a5051a2-fcf0-47f3-add1-e1e5a0d226bd/volumes" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.351911 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kzqv\" (UniqueName: \"kubernetes.io/projected/23d88b3c-33af-4075-8e30-1bca0b848ee3-kube-api-access-4kzqv\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.351997 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.352021 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.352042 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-config-data\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.352360 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-scripts\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.352537 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d88b3c-33af-4075-8e30-1bca0b848ee3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.454502 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kzqv\" (UniqueName: \"kubernetes.io/projected/23d88b3c-33af-4075-8e30-1bca0b848ee3-kube-api-access-4kzqv\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.454604 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.454626 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.454666 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-config-data\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.454735 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-scripts\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.454781 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d88b3c-33af-4075-8e30-1bca0b848ee3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.454928 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d88b3c-33af-4075-8e30-1bca0b848ee3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.459250 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.459547 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.460241 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-config-data\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.462870 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d88b3c-33af-4075-8e30-1bca0b848ee3-scripts\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.482706 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kzqv\" (UniqueName: \"kubernetes.io/projected/23d88b3c-33af-4075-8e30-1bca0b848ee3-kube-api-access-4kzqv\") pod \"cinder-scheduler-0\" (UID: \"23d88b3c-33af-4075-8e30-1bca0b848ee3\") " pod="openstack/cinder-scheduler-0" Oct 10 16:21:40 crc kubenswrapper[4788]: I1010 16:21:40.510554 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 10 16:21:41 crc kubenswrapper[4788]: I1010 16:21:41.080647 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 10 16:21:41 crc kubenswrapper[4788]: I1010 16:21:41.105895 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23d88b3c-33af-4075-8e30-1bca0b848ee3","Type":"ContainerStarted","Data":"3add6e6d0393cac87cd4dae3dfe8d7e1d171c626049af6ce414fb206d63d7e33"} Oct 10 16:21:42 crc kubenswrapper[4788]: I1010 16:21:42.135034 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23d88b3c-33af-4075-8e30-1bca0b848ee3","Type":"ContainerStarted","Data":"59a5b34eb9f34c76f6e32bf7d117b8abe10386d42b9bcdbd1249c0802a8b7606"} Oct 10 16:21:43 crc kubenswrapper[4788]: I1010 16:21:43.153745 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"23d88b3c-33af-4075-8e30-1bca0b848ee3","Type":"ContainerStarted","Data":"9c470070136ae2bb2d5e1935ef39768ab0847c8e0e991c34776e682da8e49fb3"} Oct 10 16:21:44 crc kubenswrapper[4788]: I1010 16:21:44.502584 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 10 16:21:44 crc kubenswrapper[4788]: I1010 16:21:44.542019 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.541993473 podStartE2EDuration="4.541993473s" podCreationTimestamp="2025-10-10 16:21:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:21:43.182731993 +0000 UTC m=+5805.632447551" watchObservedRunningTime="2025-10-10 16:21:44.541993473 +0000 UTC m=+5806.991709041" Oct 10 16:21:45 crc kubenswrapper[4788]: I1010 16:21:45.511621 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 10 16:21:50 crc kubenswrapper[4788]: I1010 16:21:50.238042 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:21:50 crc kubenswrapper[4788]: E1010 16:21:50.240306 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:21:50 crc kubenswrapper[4788]: I1010 16:21:50.696426 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 10 16:22:03 crc kubenswrapper[4788]: I1010 16:22:03.234128 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:22:03 crc kubenswrapper[4788]: E1010 16:22:03.234920 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:22:13 crc kubenswrapper[4788]: I1010 16:22:13.063497 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-lx4nz"] Oct 10 16:22:13 crc kubenswrapper[4788]: I1010 16:22:13.071445 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-lx4nz"] Oct 10 16:22:14 crc kubenswrapper[4788]: I1010 16:22:14.256556 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29e258a4-1339-4d09-9d34-bb75e08d761d" path="/var/lib/kubelet/pods/29e258a4-1339-4d09-9d34-bb75e08d761d/volumes" Oct 10 16:22:16 crc kubenswrapper[4788]: I1010 16:22:16.234204 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:22:16 crc kubenswrapper[4788]: E1010 16:22:16.235854 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:22:21 crc kubenswrapper[4788]: I1010 16:22:21.881528 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l7vj2"] Oct 10 16:22:21 crc kubenswrapper[4788]: I1010 16:22:21.884944 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:21 crc kubenswrapper[4788]: I1010 16:22:21.906598 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l7vj2"] Oct 10 16:22:21 crc kubenswrapper[4788]: I1010 16:22:21.985064 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7tcc\" (UniqueName: \"kubernetes.io/projected/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-kube-api-access-b7tcc\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:21 crc kubenswrapper[4788]: I1010 16:22:21.985181 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-catalog-content\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:21 crc kubenswrapper[4788]: I1010 16:22:21.985211 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-utilities\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:22 crc kubenswrapper[4788]: I1010 16:22:22.087366 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-catalog-content\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:22 crc kubenswrapper[4788]: I1010 16:22:22.087427 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-utilities\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:22 crc kubenswrapper[4788]: I1010 16:22:22.087594 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7tcc\" (UniqueName: \"kubernetes.io/projected/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-kube-api-access-b7tcc\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:22 crc kubenswrapper[4788]: I1010 16:22:22.087884 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-utilities\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:22 crc kubenswrapper[4788]: I1010 16:22:22.087884 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-catalog-content\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:22 crc kubenswrapper[4788]: I1010 16:22:22.113248 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7tcc\" (UniqueName: \"kubernetes.io/projected/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-kube-api-access-b7tcc\") pod \"certified-operators-l7vj2\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:22 crc kubenswrapper[4788]: I1010 16:22:22.219215 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:22 crc kubenswrapper[4788]: I1010 16:22:22.754690 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l7vj2"] Oct 10 16:22:23 crc kubenswrapper[4788]: I1010 16:22:23.047294 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-ec63-account-create-9gfbp"] Oct 10 16:22:23 crc kubenswrapper[4788]: I1010 16:22:23.064092 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-ec63-account-create-9gfbp"] Oct 10 16:22:23 crc kubenswrapper[4788]: I1010 16:22:23.611685 4788 generic.go:334] "Generic (PLEG): container finished" podID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerID="08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a" exitCode=0 Oct 10 16:22:23 crc kubenswrapper[4788]: I1010 16:22:23.611725 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7vj2" event={"ID":"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8","Type":"ContainerDied","Data":"08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a"} Oct 10 16:22:23 crc kubenswrapper[4788]: I1010 16:22:23.611748 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7vj2" event={"ID":"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8","Type":"ContainerStarted","Data":"596fca83cbed7dbdcdc366db7e0fb7f98959fa4b4ccce14a8fc9a40184f8ed53"} Oct 10 16:22:24 crc kubenswrapper[4788]: I1010 16:22:24.245903 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23" path="/var/lib/kubelet/pods/81e2b2a4-1ee3-4164-8f80-b1bf5ebdca23/volumes" Oct 10 16:22:25 crc kubenswrapper[4788]: I1010 16:22:25.640076 4788 generic.go:334] "Generic (PLEG): container finished" podID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerID="d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1" exitCode=0 Oct 10 16:22:25 crc kubenswrapper[4788]: I1010 16:22:25.640546 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7vj2" event={"ID":"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8","Type":"ContainerDied","Data":"d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1"} Oct 10 16:22:26 crc kubenswrapper[4788]: I1010 16:22:26.653517 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7vj2" event={"ID":"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8","Type":"ContainerStarted","Data":"0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50"} Oct 10 16:22:26 crc kubenswrapper[4788]: I1010 16:22:26.679830 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l7vj2" podStartSLOduration=3.24759006 podStartE2EDuration="5.679808972s" podCreationTimestamp="2025-10-10 16:22:21 +0000 UTC" firstStartedPulling="2025-10-10 16:22:23.614462015 +0000 UTC m=+5846.064177603" lastFinishedPulling="2025-10-10 16:22:26.046680957 +0000 UTC m=+5848.496396515" observedRunningTime="2025-10-10 16:22:26.673227914 +0000 UTC m=+5849.122943462" watchObservedRunningTime="2025-10-10 16:22:26.679808972 +0000 UTC m=+5849.129524520" Oct 10 16:22:29 crc kubenswrapper[4788]: I1010 16:22:29.234487 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:22:29 crc kubenswrapper[4788]: E1010 16:22:29.235007 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:22:30 crc kubenswrapper[4788]: I1010 16:22:30.034408 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xbzjx"] Oct 10 16:22:30 crc kubenswrapper[4788]: I1010 16:22:30.067860 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xbzjx"] Oct 10 16:22:30 crc kubenswrapper[4788]: I1010 16:22:30.247533 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="459619f5-5e06-467c-b778-3b063e727da4" path="/var/lib/kubelet/pods/459619f5-5e06-467c-b778-3b063e727da4/volumes" Oct 10 16:22:32 crc kubenswrapper[4788]: I1010 16:22:32.219553 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:32 crc kubenswrapper[4788]: I1010 16:22:32.219823 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:32 crc kubenswrapper[4788]: I1010 16:22:32.298504 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:32 crc kubenswrapper[4788]: I1010 16:22:32.763994 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:32 crc kubenswrapper[4788]: I1010 16:22:32.817002 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l7vj2"] Oct 10 16:22:34 crc kubenswrapper[4788]: I1010 16:22:34.728252 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l7vj2" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerName="registry-server" containerID="cri-o://0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50" gracePeriod=2 Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.221246 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.278420 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7tcc\" (UniqueName: \"kubernetes.io/projected/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-kube-api-access-b7tcc\") pod \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.278510 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-catalog-content\") pod \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.278612 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-utilities\") pod \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\" (UID: \"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8\") " Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.279708 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-utilities" (OuterVolumeSpecName: "utilities") pod "397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" (UID: "397188f8-d03d-45b5-8f0b-e5b0ab0c66c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.288804 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-kube-api-access-b7tcc" (OuterVolumeSpecName: "kube-api-access-b7tcc") pod "397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" (UID: "397188f8-d03d-45b5-8f0b-e5b0ab0c66c8"). InnerVolumeSpecName "kube-api-access-b7tcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.326428 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" (UID: "397188f8-d03d-45b5-8f0b-e5b0ab0c66c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.379498 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.379532 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7tcc\" (UniqueName: \"kubernetes.io/projected/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-kube-api-access-b7tcc\") on node \"crc\" DevicePath \"\"" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.379542 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.741495 4788 generic.go:334] "Generic (PLEG): container finished" podID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerID="0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50" exitCode=0 Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.741537 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7vj2" event={"ID":"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8","Type":"ContainerDied","Data":"0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50"} Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.741565 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7vj2" event={"ID":"397188f8-d03d-45b5-8f0b-e5b0ab0c66c8","Type":"ContainerDied","Data":"596fca83cbed7dbdcdc366db7e0fb7f98959fa4b4ccce14a8fc9a40184f8ed53"} Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.741581 4788 scope.go:117] "RemoveContainer" containerID="0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.741604 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l7vj2" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.772128 4788 scope.go:117] "RemoveContainer" containerID="d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.791301 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l7vj2"] Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.799310 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l7vj2"] Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.811904 4788 scope.go:117] "RemoveContainer" containerID="08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.878911 4788 scope.go:117] "RemoveContainer" containerID="0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50" Oct 10 16:22:35 crc kubenswrapper[4788]: E1010 16:22:35.879303 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50\": container with ID starting with 0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50 not found: ID does not exist" containerID="0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.879364 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50"} err="failed to get container status \"0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50\": rpc error: code = NotFound desc = could not find container \"0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50\": container with ID starting with 0a8974e4e99f486a8506fd001d4e4961a5314ca2602e2552f38f88c23924cf50 not found: ID does not exist" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.879398 4788 scope.go:117] "RemoveContainer" containerID="d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1" Oct 10 16:22:35 crc kubenswrapper[4788]: E1010 16:22:35.880020 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1\": container with ID starting with d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1 not found: ID does not exist" containerID="d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.880062 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1"} err="failed to get container status \"d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1\": rpc error: code = NotFound desc = could not find container \"d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1\": container with ID starting with d1bafdc75b58e706c6e15a7ccdb292f9cafbc9101448194c8069874ad6d06ca1 not found: ID does not exist" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.880089 4788 scope.go:117] "RemoveContainer" containerID="08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a" Oct 10 16:22:35 crc kubenswrapper[4788]: E1010 16:22:35.880563 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a\": container with ID starting with 08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a not found: ID does not exist" containerID="08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a" Oct 10 16:22:35 crc kubenswrapper[4788]: I1010 16:22:35.880596 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a"} err="failed to get container status \"08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a\": rpc error: code = NotFound desc = could not find container \"08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a\": container with ID starting with 08f61fa398b698dcf4c4ff587e271e0bd0410488f4f2e34b391674d59486e02a not found: ID does not exist" Oct 10 16:22:36 crc kubenswrapper[4788]: I1010 16:22:36.245894 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" path="/var/lib/kubelet/pods/397188f8-d03d-45b5-8f0b-e5b0ab0c66c8/volumes" Oct 10 16:22:40 crc kubenswrapper[4788]: I1010 16:22:40.234419 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:22:40 crc kubenswrapper[4788]: E1010 16:22:40.235285 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:22:43 crc kubenswrapper[4788]: I1010 16:22:43.055281 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sb82b"] Oct 10 16:22:43 crc kubenswrapper[4788]: I1010 16:22:43.068948 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sb82b"] Oct 10 16:22:44 crc kubenswrapper[4788]: I1010 16:22:44.257944 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf92e46-caea-4151-852c-7e29c99bb230" path="/var/lib/kubelet/pods/8bf92e46-caea-4151-852c-7e29c99bb230/volumes" Oct 10 16:22:54 crc kubenswrapper[4788]: I1010 16:22:54.234871 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:22:54 crc kubenswrapper[4788]: E1010 16:22:54.235662 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:23:05 crc kubenswrapper[4788]: I1010 16:23:05.235357 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:23:05 crc kubenswrapper[4788]: E1010 16:23:05.235982 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:23:13 crc kubenswrapper[4788]: I1010 16:23:13.359799 4788 scope.go:117] "RemoveContainer" containerID="2e8184017a6d3b1f51d23bc5e9b714a36c72a832e02816022fed273a2eabb69d" Oct 10 16:23:13 crc kubenswrapper[4788]: I1010 16:23:13.421708 4788 scope.go:117] "RemoveContainer" containerID="7564209e7512c76e32610dafec7a18b7d2b77184904e34abbd361ab7cbab0884" Oct 10 16:23:13 crc kubenswrapper[4788]: I1010 16:23:13.444851 4788 scope.go:117] "RemoveContainer" containerID="52b5ef897cd24c50ed474f6ccd70bcc81d26072e3b46ec8e7117f788ace6386c" Oct 10 16:23:13 crc kubenswrapper[4788]: I1010 16:23:13.524761 4788 scope.go:117] "RemoveContainer" containerID="0e3848714678ca88f78feeaef5b60ab7e762fab156a35b1fdff6b8624884102c" Oct 10 16:23:17 crc kubenswrapper[4788]: I1010 16:23:17.234555 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:23:17 crc kubenswrapper[4788]: E1010 16:23:17.235216 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:23:29 crc kubenswrapper[4788]: I1010 16:23:29.234340 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:23:29 crc kubenswrapper[4788]: E1010 16:23:29.235280 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.948943 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5k2fn"] Oct 10 16:23:32 crc kubenswrapper[4788]: E1010 16:23:32.949853 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerName="registry-server" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.949867 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerName="registry-server" Oct 10 16:23:32 crc kubenswrapper[4788]: E1010 16:23:32.949899 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerName="extract-content" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.949905 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerName="extract-content" Oct 10 16:23:32 crc kubenswrapper[4788]: E1010 16:23:32.949923 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerName="extract-utilities" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.949931 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerName="extract-utilities" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.950101 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="397188f8-d03d-45b5-8f0b-e5b0ab0c66c8" containerName="registry-server" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.950807 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.953175 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.954695 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-tjnjx" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.962870 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5k2fn"] Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.973549 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-t7sxw"] Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.975766 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.981392 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-run-ovn\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.981765 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-run\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.981942 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m6lh\" (UniqueName: \"kubernetes.io/projected/b7674601-b383-40b8-80ad-0864b4114d70-kube-api-access-4m6lh\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.982070 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-log-ovn\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.982269 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7674601-b383-40b8-80ad-0864b4114d70-scripts\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:32 crc kubenswrapper[4788]: I1010 16:23:32.987885 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-t7sxw"] Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084216 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-lib\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084284 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m6lh\" (UniqueName: \"kubernetes.io/projected/b7674601-b383-40b8-80ad-0864b4114d70-kube-api-access-4m6lh\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084304 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2b7ee70-2711-4155-a154-1fe0b96984e5-scripts\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084336 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-log-ovn\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084372 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7674601-b383-40b8-80ad-0864b4114d70-scripts\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084390 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-etc-ovs\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084412 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-log\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084492 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vd2x\" (UniqueName: \"kubernetes.io/projected/a2b7ee70-2711-4155-a154-1fe0b96984e5-kube-api-access-2vd2x\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084587 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-run-ovn\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084653 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-run\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084711 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-run\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084714 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-log-ovn\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084749 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-run-ovn\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.084826 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b7674601-b383-40b8-80ad-0864b4114d70-var-run\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.086814 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7674601-b383-40b8-80ad-0864b4114d70-scripts\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.106668 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m6lh\" (UniqueName: \"kubernetes.io/projected/b7674601-b383-40b8-80ad-0864b4114d70-kube-api-access-4m6lh\") pod \"ovn-controller-5k2fn\" (UID: \"b7674601-b383-40b8-80ad-0864b4114d70\") " pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.185904 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-etc-ovs\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.185959 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-log\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.185995 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vd2x\" (UniqueName: \"kubernetes.io/projected/a2b7ee70-2711-4155-a154-1fe0b96984e5-kube-api-access-2vd2x\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.186098 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-run\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.186122 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-lib\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.186161 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2b7ee70-2711-4155-a154-1fe0b96984e5-scripts\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.186207 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-lib\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.186163 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-log\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.186220 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-var-run\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.186288 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/a2b7ee70-2711-4155-a154-1fe0b96984e5-etc-ovs\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.189187 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a2b7ee70-2711-4155-a154-1fe0b96984e5-scripts\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.205058 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vd2x\" (UniqueName: \"kubernetes.io/projected/a2b7ee70-2711-4155-a154-1fe0b96984e5-kube-api-access-2vd2x\") pod \"ovn-controller-ovs-t7sxw\" (UID: \"a2b7ee70-2711-4155-a154-1fe0b96984e5\") " pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.270431 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.299515 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:33 crc kubenswrapper[4788]: I1010 16:23:33.851192 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5k2fn"] Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.099075 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-t7sxw"] Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.344632 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5k2fn" event={"ID":"b7674601-b383-40b8-80ad-0864b4114d70","Type":"ContainerStarted","Data":"b6bed93e911618bb41d2f46c8a9130fa28dfa4c098bbcde0d2a2d4dcf2b1c50d"} Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.344999 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5k2fn" event={"ID":"b7674601-b383-40b8-80ad-0864b4114d70","Type":"ContainerStarted","Data":"47fed4a71b0c5983babaa8cd18ff891d3d097adc0e67d513498c117eaf84bd88"} Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.345054 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-5k2fn" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.346704 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7sxw" event={"ID":"a2b7ee70-2711-4155-a154-1fe0b96984e5","Type":"ContainerStarted","Data":"ad52d450be525f6aac486504283bc726587ab9510787ef0417262460003ec756"} Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.365736 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-5k2fn" podStartSLOduration=2.365719139 podStartE2EDuration="2.365719139s" podCreationTimestamp="2025-10-10 16:23:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:23:34.357178269 +0000 UTC m=+5916.806893827" watchObservedRunningTime="2025-10-10 16:23:34.365719139 +0000 UTC m=+5916.815434687" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.608944 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-pntng"] Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.613173 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.615785 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.625367 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-pntng"] Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.711547 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4e9e9129-a513-4255-9561-6a5f268193f2-ovs-rundir\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.711594 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4e9e9129-a513-4255-9561-6a5f268193f2-ovn-rundir\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.711631 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmrk9\" (UniqueName: \"kubernetes.io/projected/4e9e9129-a513-4255-9561-6a5f268193f2-kube-api-access-xmrk9\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.711661 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e9e9129-a513-4255-9561-6a5f268193f2-config\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.813712 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4e9e9129-a513-4255-9561-6a5f268193f2-ovs-rundir\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.813761 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4e9e9129-a513-4255-9561-6a5f268193f2-ovn-rundir\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.813795 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmrk9\" (UniqueName: \"kubernetes.io/projected/4e9e9129-a513-4255-9561-6a5f268193f2-kube-api-access-xmrk9\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.813819 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e9e9129-a513-4255-9561-6a5f268193f2-config\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.814696 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4e9e9129-a513-4255-9561-6a5f268193f2-ovn-rundir\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.814739 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4e9e9129-a513-4255-9561-6a5f268193f2-ovs-rundir\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.814848 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e9e9129-a513-4255-9561-6a5f268193f2-config\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.833311 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmrk9\" (UniqueName: \"kubernetes.io/projected/4e9e9129-a513-4255-9561-6a5f268193f2-kube-api-access-xmrk9\") pod \"ovn-controller-metrics-pntng\" (UID: \"4e9e9129-a513-4255-9561-6a5f268193f2\") " pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:34 crc kubenswrapper[4788]: I1010 16:23:34.946337 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-pntng" Oct 10 16:23:35 crc kubenswrapper[4788]: I1010 16:23:35.356629 4788 generic.go:334] "Generic (PLEG): container finished" podID="a2b7ee70-2711-4155-a154-1fe0b96984e5" containerID="192e83781bd0a8e1b5cec2f6f6a8635e59fb2224fa47f1b8a5f71086c937f8fd" exitCode=0 Oct 10 16:23:35 crc kubenswrapper[4788]: I1010 16:23:35.356727 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7sxw" event={"ID":"a2b7ee70-2711-4155-a154-1fe0b96984e5","Type":"ContainerDied","Data":"192e83781bd0a8e1b5cec2f6f6a8635e59fb2224fa47f1b8a5f71086c937f8fd"} Oct 10 16:23:35 crc kubenswrapper[4788]: I1010 16:23:35.410768 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-pntng"] Oct 10 16:23:35 crc kubenswrapper[4788]: W1010 16:23:35.423492 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e9e9129_a513_4255_9561_6a5f268193f2.slice/crio-f61b375cde1178ec8f9a6955a8903b319823f41c7821c83e10bbcaead6b1129a WatchSource:0}: Error finding container f61b375cde1178ec8f9a6955a8903b319823f41c7821c83e10bbcaead6b1129a: Status 404 returned error can't find the container with id f61b375cde1178ec8f9a6955a8903b319823f41c7821c83e10bbcaead6b1129a Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.079654 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-hnbp6"] Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.082083 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-hnbp6" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.096431 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-hnbp6"] Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.150653 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rsg7\" (UniqueName: \"kubernetes.io/projected/36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab-kube-api-access-7rsg7\") pod \"octavia-db-create-hnbp6\" (UID: \"36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab\") " pod="openstack/octavia-db-create-hnbp6" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.252278 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rsg7\" (UniqueName: \"kubernetes.io/projected/36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab-kube-api-access-7rsg7\") pod \"octavia-db-create-hnbp6\" (UID: \"36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab\") " pod="openstack/octavia-db-create-hnbp6" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.276887 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rsg7\" (UniqueName: \"kubernetes.io/projected/36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab-kube-api-access-7rsg7\") pod \"octavia-db-create-hnbp6\" (UID: \"36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab\") " pod="openstack/octavia-db-create-hnbp6" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.367300 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-pntng" event={"ID":"4e9e9129-a513-4255-9561-6a5f268193f2","Type":"ContainerStarted","Data":"de3aff6cbf0476aa0636ee9f1626629203bc9956ccb43ebb2fbe7ef151ad5c99"} Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.369855 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-pntng" event={"ID":"4e9e9129-a513-4255-9561-6a5f268193f2","Type":"ContainerStarted","Data":"f61b375cde1178ec8f9a6955a8903b319823f41c7821c83e10bbcaead6b1129a"} Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.372656 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7sxw" event={"ID":"a2b7ee70-2711-4155-a154-1fe0b96984e5","Type":"ContainerStarted","Data":"e972b8739c1c1fab96f1710ea4b5f0e413c15153b4b5261a8341601ad72694b3"} Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.372840 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7sxw" event={"ID":"a2b7ee70-2711-4155-a154-1fe0b96984e5","Type":"ContainerStarted","Data":"4ffbf4d74b17df63f291f57f5c12e4a66c587a970ed46ce1dfab41e266f62692"} Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.373290 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.373403 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.389622 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-pntng" podStartSLOduration=2.389602852 podStartE2EDuration="2.389602852s" podCreationTimestamp="2025-10-10 16:23:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:23:36.384143381 +0000 UTC m=+5918.833858929" watchObservedRunningTime="2025-10-10 16:23:36.389602852 +0000 UTC m=+5918.839318410" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.401798 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-hnbp6" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.414773 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-t7sxw" podStartSLOduration=4.414755788 podStartE2EDuration="4.414755788s" podCreationTimestamp="2025-10-10 16:23:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:23:36.409770339 +0000 UTC m=+5918.859485887" watchObservedRunningTime="2025-10-10 16:23:36.414755788 +0000 UTC m=+5918.864471336" Oct 10 16:23:36 crc kubenswrapper[4788]: I1010 16:23:36.905073 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-hnbp6"] Oct 10 16:23:37 crc kubenswrapper[4788]: I1010 16:23:37.382754 4788 generic.go:334] "Generic (PLEG): container finished" podID="36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab" containerID="9f166804d7756e8bfe9c780cc68371874fab9f8a5c1ac77e3d595d81fd1b45ee" exitCode=0 Oct 10 16:23:37 crc kubenswrapper[4788]: I1010 16:23:37.383177 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-hnbp6" event={"ID":"36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab","Type":"ContainerDied","Data":"9f166804d7756e8bfe9c780cc68371874fab9f8a5c1ac77e3d595d81fd1b45ee"} Oct 10 16:23:37 crc kubenswrapper[4788]: I1010 16:23:37.383203 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-hnbp6" event={"ID":"36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab","Type":"ContainerStarted","Data":"2e2e1d1a0ba63472e50aed7609d705401f10282348d78da7b5a9a534b76da1d6"} Oct 10 16:23:38 crc kubenswrapper[4788]: I1010 16:23:38.749394 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-hnbp6" Oct 10 16:23:38 crc kubenswrapper[4788]: I1010 16:23:38.801594 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rsg7\" (UniqueName: \"kubernetes.io/projected/36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab-kube-api-access-7rsg7\") pod \"36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab\" (UID: \"36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab\") " Oct 10 16:23:38 crc kubenswrapper[4788]: I1010 16:23:38.809116 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab-kube-api-access-7rsg7" (OuterVolumeSpecName: "kube-api-access-7rsg7") pod "36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab" (UID: "36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab"). InnerVolumeSpecName "kube-api-access-7rsg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:23:38 crc kubenswrapper[4788]: I1010 16:23:38.904645 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rsg7\" (UniqueName: \"kubernetes.io/projected/36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab-kube-api-access-7rsg7\") on node \"crc\" DevicePath \"\"" Oct 10 16:23:39 crc kubenswrapper[4788]: I1010 16:23:39.408157 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-hnbp6" event={"ID":"36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab","Type":"ContainerDied","Data":"2e2e1d1a0ba63472e50aed7609d705401f10282348d78da7b5a9a534b76da1d6"} Oct 10 16:23:39 crc kubenswrapper[4788]: I1010 16:23:39.408206 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e2e1d1a0ba63472e50aed7609d705401f10282348d78da7b5a9a534b76da1d6" Oct 10 16:23:39 crc kubenswrapper[4788]: I1010 16:23:39.408275 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-hnbp6" Oct 10 16:23:43 crc kubenswrapper[4788]: I1010 16:23:43.234214 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:23:44 crc kubenswrapper[4788]: I1010 16:23:44.474205 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"e510a005acb06733b8f662e2f362c13e2ba92b9f626ca5375903dde617f1f2fe"} Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.089261 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-725d-account-create-h76lw"] Oct 10 16:23:48 crc kubenswrapper[4788]: E1010 16:23:48.090339 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab" containerName="mariadb-database-create" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.090355 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab" containerName="mariadb-database-create" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.090619 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab" containerName="mariadb-database-create" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.091452 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-725d-account-create-h76lw" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.095852 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.097654 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-725d-account-create-h76lw"] Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.148856 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4hhj\" (UniqueName: \"kubernetes.io/projected/b437783d-adf5-442e-ba13-3f2abd872426-kube-api-access-z4hhj\") pod \"octavia-725d-account-create-h76lw\" (UID: \"b437783d-adf5-442e-ba13-3f2abd872426\") " pod="openstack/octavia-725d-account-create-h76lw" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.250654 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4hhj\" (UniqueName: \"kubernetes.io/projected/b437783d-adf5-442e-ba13-3f2abd872426-kube-api-access-z4hhj\") pod \"octavia-725d-account-create-h76lw\" (UID: \"b437783d-adf5-442e-ba13-3f2abd872426\") " pod="openstack/octavia-725d-account-create-h76lw" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.273851 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4hhj\" (UniqueName: \"kubernetes.io/projected/b437783d-adf5-442e-ba13-3f2abd872426-kube-api-access-z4hhj\") pod \"octavia-725d-account-create-h76lw\" (UID: \"b437783d-adf5-442e-ba13-3f2abd872426\") " pod="openstack/octavia-725d-account-create-h76lw" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.472851 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-725d-account-create-h76lw" Oct 10 16:23:48 crc kubenswrapper[4788]: I1010 16:23:48.926244 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-725d-account-create-h76lw"] Oct 10 16:23:49 crc kubenswrapper[4788]: I1010 16:23:49.534762 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-725d-account-create-h76lw" event={"ID":"b437783d-adf5-442e-ba13-3f2abd872426","Type":"ContainerDied","Data":"7efe00d56a6fbeabf4a56a72fe9e1377374e2723ed9c354311e0d50558d5a5de"} Oct 10 16:23:49 crc kubenswrapper[4788]: I1010 16:23:49.534591 4788 generic.go:334] "Generic (PLEG): container finished" podID="b437783d-adf5-442e-ba13-3f2abd872426" containerID="7efe00d56a6fbeabf4a56a72fe9e1377374e2723ed9c354311e0d50558d5a5de" exitCode=0 Oct 10 16:23:49 crc kubenswrapper[4788]: I1010 16:23:49.535369 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-725d-account-create-h76lw" event={"ID":"b437783d-adf5-442e-ba13-3f2abd872426","Type":"ContainerStarted","Data":"1667adcffa29c75552d7cfb79e33d425bd99b15645d6109f6bdaaa2decb1357a"} Oct 10 16:23:50 crc kubenswrapper[4788]: I1010 16:23:50.959076 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-725d-account-create-h76lw" Oct 10 16:23:51 crc kubenswrapper[4788]: I1010 16:23:51.113525 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4hhj\" (UniqueName: \"kubernetes.io/projected/b437783d-adf5-442e-ba13-3f2abd872426-kube-api-access-z4hhj\") pod \"b437783d-adf5-442e-ba13-3f2abd872426\" (UID: \"b437783d-adf5-442e-ba13-3f2abd872426\") " Oct 10 16:23:51 crc kubenswrapper[4788]: I1010 16:23:51.124750 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b437783d-adf5-442e-ba13-3f2abd872426-kube-api-access-z4hhj" (OuterVolumeSpecName: "kube-api-access-z4hhj") pod "b437783d-adf5-442e-ba13-3f2abd872426" (UID: "b437783d-adf5-442e-ba13-3f2abd872426"). InnerVolumeSpecName "kube-api-access-z4hhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:23:51 crc kubenswrapper[4788]: I1010 16:23:51.216683 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4hhj\" (UniqueName: \"kubernetes.io/projected/b437783d-adf5-442e-ba13-3f2abd872426-kube-api-access-z4hhj\") on node \"crc\" DevicePath \"\"" Oct 10 16:23:51 crc kubenswrapper[4788]: I1010 16:23:51.559997 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-725d-account-create-h76lw" event={"ID":"b437783d-adf5-442e-ba13-3f2abd872426","Type":"ContainerDied","Data":"1667adcffa29c75552d7cfb79e33d425bd99b15645d6109f6bdaaa2decb1357a"} Oct 10 16:23:51 crc kubenswrapper[4788]: I1010 16:23:51.560087 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1667adcffa29c75552d7cfb79e33d425bd99b15645d6109f6bdaaa2decb1357a" Oct 10 16:23:51 crc kubenswrapper[4788]: I1010 16:23:51.560013 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-725d-account-create-h76lw" Oct 10 16:23:54 crc kubenswrapper[4788]: I1010 16:23:54.865007 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-qwtnc"] Oct 10 16:23:54 crc kubenswrapper[4788]: E1010 16:23:54.866846 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b437783d-adf5-442e-ba13-3f2abd872426" containerName="mariadb-account-create" Oct 10 16:23:54 crc kubenswrapper[4788]: I1010 16:23:54.866872 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b437783d-adf5-442e-ba13-3f2abd872426" containerName="mariadb-account-create" Oct 10 16:23:54 crc kubenswrapper[4788]: I1010 16:23:54.867240 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b437783d-adf5-442e-ba13-3f2abd872426" containerName="mariadb-account-create" Oct 10 16:23:54 crc kubenswrapper[4788]: I1010 16:23:54.868225 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-qwtnc" Oct 10 16:23:54 crc kubenswrapper[4788]: I1010 16:23:54.891613 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-qwtnc"] Oct 10 16:23:55 crc kubenswrapper[4788]: I1010 16:23:55.037264 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcfwf\" (UniqueName: \"kubernetes.io/projected/25598815-a0ed-4d38-9e06-89bac3efd642-kube-api-access-tcfwf\") pod \"octavia-persistence-db-create-qwtnc\" (UID: \"25598815-a0ed-4d38-9e06-89bac3efd642\") " pod="openstack/octavia-persistence-db-create-qwtnc" Oct 10 16:23:55 crc kubenswrapper[4788]: I1010 16:23:55.148299 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcfwf\" (UniqueName: \"kubernetes.io/projected/25598815-a0ed-4d38-9e06-89bac3efd642-kube-api-access-tcfwf\") pod \"octavia-persistence-db-create-qwtnc\" (UID: \"25598815-a0ed-4d38-9e06-89bac3efd642\") " pod="openstack/octavia-persistence-db-create-qwtnc" Oct 10 16:23:55 crc kubenswrapper[4788]: I1010 16:23:55.174009 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcfwf\" (UniqueName: \"kubernetes.io/projected/25598815-a0ed-4d38-9e06-89bac3efd642-kube-api-access-tcfwf\") pod \"octavia-persistence-db-create-qwtnc\" (UID: \"25598815-a0ed-4d38-9e06-89bac3efd642\") " pod="openstack/octavia-persistence-db-create-qwtnc" Oct 10 16:23:55 crc kubenswrapper[4788]: I1010 16:23:55.194437 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-qwtnc" Oct 10 16:23:55 crc kubenswrapper[4788]: I1010 16:23:55.679457 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-qwtnc"] Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.625440 4788 generic.go:334] "Generic (PLEG): container finished" podID="25598815-a0ed-4d38-9e06-89bac3efd642" containerID="5b5bd771151fef05295f13aef1233a50d12abbc8a7b111e015e9bfaf563aa042" exitCode=0 Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.625548 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-qwtnc" event={"ID":"25598815-a0ed-4d38-9e06-89bac3efd642","Type":"ContainerDied","Data":"5b5bd771151fef05295f13aef1233a50d12abbc8a7b111e015e9bfaf563aa042"} Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.625887 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-qwtnc" event={"ID":"25598815-a0ed-4d38-9e06-89bac3efd642","Type":"ContainerStarted","Data":"cf25a103e45f832657ff39f62858354c6f267b30cb498752431446348c4408e7"} Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.880927 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-62b9s"] Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.883585 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.894815 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-62b9s"] Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.994253 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-utilities\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.994888 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g69q6\" (UniqueName: \"kubernetes.io/projected/d277f452-3e5f-4cb1-b511-bb79b496fc10-kube-api-access-g69q6\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:56 crc kubenswrapper[4788]: I1010 16:23:56.994954 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-catalog-content\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.096548 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-utilities\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.096683 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g69q6\" (UniqueName: \"kubernetes.io/projected/d277f452-3e5f-4cb1-b511-bb79b496fc10-kube-api-access-g69q6\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.096711 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-catalog-content\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.097315 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-utilities\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.097341 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-catalog-content\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.117124 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g69q6\" (UniqueName: \"kubernetes.io/projected/d277f452-3e5f-4cb1-b511-bb79b496fc10-kube-api-access-g69q6\") pod \"redhat-operators-62b9s\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.215358 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.677601 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-62b9s"] Oct 10 16:23:57 crc kubenswrapper[4788]: W1010 16:23:57.684485 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd277f452_3e5f_4cb1_b511_bb79b496fc10.slice/crio-aeef38f22ee7a944492f99a46d5aae7e3ad0aa328a0c182e99f54405e2c9259e WatchSource:0}: Error finding container aeef38f22ee7a944492f99a46d5aae7e3ad0aa328a0c182e99f54405e2c9259e: Status 404 returned error can't find the container with id aeef38f22ee7a944492f99a46d5aae7e3ad0aa328a0c182e99f54405e2c9259e Oct 10 16:23:57 crc kubenswrapper[4788]: I1010 16:23:57.990310 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-qwtnc" Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.115997 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcfwf\" (UniqueName: \"kubernetes.io/projected/25598815-a0ed-4d38-9e06-89bac3efd642-kube-api-access-tcfwf\") pod \"25598815-a0ed-4d38-9e06-89bac3efd642\" (UID: \"25598815-a0ed-4d38-9e06-89bac3efd642\") " Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.122502 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25598815-a0ed-4d38-9e06-89bac3efd642-kube-api-access-tcfwf" (OuterVolumeSpecName: "kube-api-access-tcfwf") pod "25598815-a0ed-4d38-9e06-89bac3efd642" (UID: "25598815-a0ed-4d38-9e06-89bac3efd642"). InnerVolumeSpecName "kube-api-access-tcfwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.218299 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcfwf\" (UniqueName: \"kubernetes.io/projected/25598815-a0ed-4d38-9e06-89bac3efd642-kube-api-access-tcfwf\") on node \"crc\" DevicePath \"\"" Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.665620 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-qwtnc" event={"ID":"25598815-a0ed-4d38-9e06-89bac3efd642","Type":"ContainerDied","Data":"cf25a103e45f832657ff39f62858354c6f267b30cb498752431446348c4408e7"} Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.665687 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf25a103e45f832657ff39f62858354c6f267b30cb498752431446348c4408e7" Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.665551 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-qwtnc" Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.673431 4788 generic.go:334] "Generic (PLEG): container finished" podID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerID="0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7" exitCode=0 Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.673504 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62b9s" event={"ID":"d277f452-3e5f-4cb1-b511-bb79b496fc10","Type":"ContainerDied","Data":"0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7"} Oct 10 16:23:58 crc kubenswrapper[4788]: I1010 16:23:58.673741 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62b9s" event={"ID":"d277f452-3e5f-4cb1-b511-bb79b496fc10","Type":"ContainerStarted","Data":"aeef38f22ee7a944492f99a46d5aae7e3ad0aa328a0c182e99f54405e2c9259e"} Oct 10 16:23:59 crc kubenswrapper[4788]: I1010 16:23:59.686923 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62b9s" event={"ID":"d277f452-3e5f-4cb1-b511-bb79b496fc10","Type":"ContainerStarted","Data":"1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb"} Oct 10 16:24:00 crc kubenswrapper[4788]: I1010 16:24:00.703936 4788 generic.go:334] "Generic (PLEG): container finished" podID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerID="1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb" exitCode=0 Oct 10 16:24:00 crc kubenswrapper[4788]: I1010 16:24:00.704044 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62b9s" event={"ID":"d277f452-3e5f-4cb1-b511-bb79b496fc10","Type":"ContainerDied","Data":"1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb"} Oct 10 16:24:01 crc kubenswrapper[4788]: I1010 16:24:01.716027 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62b9s" event={"ID":"d277f452-3e5f-4cb1-b511-bb79b496fc10","Type":"ContainerStarted","Data":"19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20"} Oct 10 16:24:01 crc kubenswrapper[4788]: I1010 16:24:01.736955 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-62b9s" podStartSLOduration=3.277166457 podStartE2EDuration="5.736935158s" podCreationTimestamp="2025-10-10 16:23:56 +0000 UTC" firstStartedPulling="2025-10-10 16:23:58.677310337 +0000 UTC m=+5941.127025885" lastFinishedPulling="2025-10-10 16:24:01.137079028 +0000 UTC m=+5943.586794586" observedRunningTime="2025-10-10 16:24:01.734828465 +0000 UTC m=+5944.184544003" watchObservedRunningTime="2025-10-10 16:24:01.736935158 +0000 UTC m=+5944.186650706" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.037503 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-1c0e-account-create-lmchj"] Oct 10 16:24:06 crc kubenswrapper[4788]: E1010 16:24:06.038358 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25598815-a0ed-4d38-9e06-89bac3efd642" containerName="mariadb-database-create" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.038372 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="25598815-a0ed-4d38-9e06-89bac3efd642" containerName="mariadb-database-create" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.038583 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="25598815-a0ed-4d38-9e06-89bac3efd642" containerName="mariadb-database-create" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.039169 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1c0e-account-create-lmchj" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.044215 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.061637 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1c0e-account-create-lmchj"] Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.213911 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5768f\" (UniqueName: \"kubernetes.io/projected/fd8c72b5-9106-40ec-9b4e-6d993a0799c5-kube-api-access-5768f\") pod \"octavia-1c0e-account-create-lmchj\" (UID: \"fd8c72b5-9106-40ec-9b4e-6d993a0799c5\") " pod="openstack/octavia-1c0e-account-create-lmchj" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.316317 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5768f\" (UniqueName: \"kubernetes.io/projected/fd8c72b5-9106-40ec-9b4e-6d993a0799c5-kube-api-access-5768f\") pod \"octavia-1c0e-account-create-lmchj\" (UID: \"fd8c72b5-9106-40ec-9b4e-6d993a0799c5\") " pod="openstack/octavia-1c0e-account-create-lmchj" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.348558 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5768f\" (UniqueName: \"kubernetes.io/projected/fd8c72b5-9106-40ec-9b4e-6d993a0799c5-kube-api-access-5768f\") pod \"octavia-1c0e-account-create-lmchj\" (UID: \"fd8c72b5-9106-40ec-9b4e-6d993a0799c5\") " pod="openstack/octavia-1c0e-account-create-lmchj" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.371607 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1c0e-account-create-lmchj" Oct 10 16:24:06 crc kubenswrapper[4788]: I1010 16:24:06.871048 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1c0e-account-create-lmchj"] Oct 10 16:24:07 crc kubenswrapper[4788]: I1010 16:24:07.217233 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:24:07 crc kubenswrapper[4788]: I1010 16:24:07.217270 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:24:07 crc kubenswrapper[4788]: I1010 16:24:07.288336 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:24:07 crc kubenswrapper[4788]: I1010 16:24:07.781087 4788 generic.go:334] "Generic (PLEG): container finished" podID="fd8c72b5-9106-40ec-9b4e-6d993a0799c5" containerID="39683b019f3f5c7bc838991615f91a9f353885734145b9ed3cec5c378499a95c" exitCode=0 Oct 10 16:24:07 crc kubenswrapper[4788]: I1010 16:24:07.781181 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1c0e-account-create-lmchj" event={"ID":"fd8c72b5-9106-40ec-9b4e-6d993a0799c5","Type":"ContainerDied","Data":"39683b019f3f5c7bc838991615f91a9f353885734145b9ed3cec5c378499a95c"} Oct 10 16:24:07 crc kubenswrapper[4788]: I1010 16:24:07.781268 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1c0e-account-create-lmchj" event={"ID":"fd8c72b5-9106-40ec-9b4e-6d993a0799c5","Type":"ContainerStarted","Data":"5f1ffc06474e2f8e3a52b5d99bcab8aa0ee24ff46d58051f8436ec4819af54e5"} Oct 10 16:24:07 crc kubenswrapper[4788]: I1010 16:24:07.834932 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:24:07 crc kubenswrapper[4788]: I1010 16:24:07.890567 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-62b9s"] Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.312437 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5k2fn" podUID="b7674601-b383-40b8-80ad-0864b4114d70" containerName="ovn-controller" probeResult="failure" output=< Oct 10 16:24:08 crc kubenswrapper[4788]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 10 16:24:08 crc kubenswrapper[4788]: > Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.357933 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.367848 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-t7sxw" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.515020 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5k2fn-config-w7ljc"] Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.517122 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.520422 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.534909 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5k2fn-config-w7ljc"] Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.668564 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.668704 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf9t6\" (UniqueName: \"kubernetes.io/projected/6a34bec5-c302-4e14-b422-669bba4d2ffb-kube-api-access-tf9t6\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.668915 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-scripts\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.669070 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-log-ovn\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.669125 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run-ovn\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.669183 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-additional-scripts\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771238 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771323 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf9t6\" (UniqueName: \"kubernetes.io/projected/6a34bec5-c302-4e14-b422-669bba4d2ffb-kube-api-access-tf9t6\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771379 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-scripts\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771438 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-log-ovn\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771482 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run-ovn\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771514 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-additional-scripts\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771686 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run-ovn\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.771643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-log-ovn\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.772407 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-additional-scripts\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.773836 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-scripts\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.795856 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf9t6\" (UniqueName: \"kubernetes.io/projected/6a34bec5-c302-4e14-b422-669bba4d2ffb-kube-api-access-tf9t6\") pod \"ovn-controller-5k2fn-config-w7ljc\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:08 crc kubenswrapper[4788]: I1010 16:24:08.837776 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.145931 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1c0e-account-create-lmchj" Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.280939 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5768f\" (UniqueName: \"kubernetes.io/projected/fd8c72b5-9106-40ec-9b4e-6d993a0799c5-kube-api-access-5768f\") pod \"fd8c72b5-9106-40ec-9b4e-6d993a0799c5\" (UID: \"fd8c72b5-9106-40ec-9b4e-6d993a0799c5\") " Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.286666 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8c72b5-9106-40ec-9b4e-6d993a0799c5-kube-api-access-5768f" (OuterVolumeSpecName: "kube-api-access-5768f") pod "fd8c72b5-9106-40ec-9b4e-6d993a0799c5" (UID: "fd8c72b5-9106-40ec-9b4e-6d993a0799c5"). InnerVolumeSpecName "kube-api-access-5768f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.348211 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5k2fn-config-w7ljc"] Oct 10 16:24:09 crc kubenswrapper[4788]: W1010 16:24:09.352395 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a34bec5_c302_4e14_b422_669bba4d2ffb.slice/crio-a24a3bc2dab3969986f5fe648d671c4897a69e892f5777d42356ae4bd98d71bd WatchSource:0}: Error finding container a24a3bc2dab3969986f5fe648d671c4897a69e892f5777d42356ae4bd98d71bd: Status 404 returned error can't find the container with id a24a3bc2dab3969986f5fe648d671c4897a69e892f5777d42356ae4bd98d71bd Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.384514 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5768f\" (UniqueName: \"kubernetes.io/projected/fd8c72b5-9106-40ec-9b4e-6d993a0799c5-kube-api-access-5768f\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.808228 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1c0e-account-create-lmchj" event={"ID":"fd8c72b5-9106-40ec-9b4e-6d993a0799c5","Type":"ContainerDied","Data":"5f1ffc06474e2f8e3a52b5d99bcab8aa0ee24ff46d58051f8436ec4819af54e5"} Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.808637 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f1ffc06474e2f8e3a52b5d99bcab8aa0ee24ff46d58051f8436ec4819af54e5" Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.808283 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1c0e-account-create-lmchj" Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.810112 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-62b9s" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerName="registry-server" containerID="cri-o://19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20" gracePeriod=2 Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.811306 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5k2fn-config-w7ljc" event={"ID":"6a34bec5-c302-4e14-b422-669bba4d2ffb","Type":"ContainerStarted","Data":"6634e9f5cb7db0dfb8b5eeb1de0480036835200f9e1d838d2abf7b09b7bed166"} Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.811342 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5k2fn-config-w7ljc" event={"ID":"6a34bec5-c302-4e14-b422-669bba4d2ffb","Type":"ContainerStarted","Data":"a24a3bc2dab3969986f5fe648d671c4897a69e892f5777d42356ae4bd98d71bd"} Oct 10 16:24:09 crc kubenswrapper[4788]: I1010 16:24:09.833629 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-5k2fn-config-w7ljc" podStartSLOduration=1.83360018 podStartE2EDuration="1.83360018s" podCreationTimestamp="2025-10-10 16:24:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:24:09.831044694 +0000 UTC m=+5952.280760242" watchObservedRunningTime="2025-10-10 16:24:09.83360018 +0000 UTC m=+5952.283315728" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.279135 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.403961 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-utilities\") pod \"d277f452-3e5f-4cb1-b511-bb79b496fc10\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.404657 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-catalog-content\") pod \"d277f452-3e5f-4cb1-b511-bb79b496fc10\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.404723 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g69q6\" (UniqueName: \"kubernetes.io/projected/d277f452-3e5f-4cb1-b511-bb79b496fc10-kube-api-access-g69q6\") pod \"d277f452-3e5f-4cb1-b511-bb79b496fc10\" (UID: \"d277f452-3e5f-4cb1-b511-bb79b496fc10\") " Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.405236 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-utilities" (OuterVolumeSpecName: "utilities") pod "d277f452-3e5f-4cb1-b511-bb79b496fc10" (UID: "d277f452-3e5f-4cb1-b511-bb79b496fc10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.405835 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.410113 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d277f452-3e5f-4cb1-b511-bb79b496fc10-kube-api-access-g69q6" (OuterVolumeSpecName: "kube-api-access-g69q6") pod "d277f452-3e5f-4cb1-b511-bb79b496fc10" (UID: "d277f452-3e5f-4cb1-b511-bb79b496fc10"). InnerVolumeSpecName "kube-api-access-g69q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.482798 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d277f452-3e5f-4cb1-b511-bb79b496fc10" (UID: "d277f452-3e5f-4cb1-b511-bb79b496fc10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.507982 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d277f452-3e5f-4cb1-b511-bb79b496fc10-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.508017 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g69q6\" (UniqueName: \"kubernetes.io/projected/d277f452-3e5f-4cb1-b511-bb79b496fc10-kube-api-access-g69q6\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.819654 4788 generic.go:334] "Generic (PLEG): container finished" podID="6a34bec5-c302-4e14-b422-669bba4d2ffb" containerID="6634e9f5cb7db0dfb8b5eeb1de0480036835200f9e1d838d2abf7b09b7bed166" exitCode=0 Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.819733 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5k2fn-config-w7ljc" event={"ID":"6a34bec5-c302-4e14-b422-669bba4d2ffb","Type":"ContainerDied","Data":"6634e9f5cb7db0dfb8b5eeb1de0480036835200f9e1d838d2abf7b09b7bed166"} Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.822170 4788 generic.go:334] "Generic (PLEG): container finished" podID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerID="19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20" exitCode=0 Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.822227 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-62b9s" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.822227 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62b9s" event={"ID":"d277f452-3e5f-4cb1-b511-bb79b496fc10","Type":"ContainerDied","Data":"19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20"} Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.822374 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-62b9s" event={"ID":"d277f452-3e5f-4cb1-b511-bb79b496fc10","Type":"ContainerDied","Data":"aeef38f22ee7a944492f99a46d5aae7e3ad0aa328a0c182e99f54405e2c9259e"} Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.822423 4788 scope.go:117] "RemoveContainer" containerID="19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.863328 4788 scope.go:117] "RemoveContainer" containerID="1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.929687 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-62b9s"] Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.939324 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-62b9s"] Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.939943 4788 scope.go:117] "RemoveContainer" containerID="0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.978619 4788 scope.go:117] "RemoveContainer" containerID="19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20" Oct 10 16:24:10 crc kubenswrapper[4788]: E1010 16:24:10.979362 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20\": container with ID starting with 19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20 not found: ID does not exist" containerID="19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.979407 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20"} err="failed to get container status \"19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20\": rpc error: code = NotFound desc = could not find container \"19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20\": container with ID starting with 19f24d850d538e1ac45642683e6b18faf28020d565f4655af2663c78749aed20 not found: ID does not exist" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.979432 4788 scope.go:117] "RemoveContainer" containerID="1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb" Oct 10 16:24:10 crc kubenswrapper[4788]: E1010 16:24:10.979769 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb\": container with ID starting with 1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb not found: ID does not exist" containerID="1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.979822 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb"} err="failed to get container status \"1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb\": rpc error: code = NotFound desc = could not find container \"1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb\": container with ID starting with 1fb8e04a6e11113f518c0c3358455c4791a3c6b84bc36e6f6703341817bcb7cb not found: ID does not exist" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.979859 4788 scope.go:117] "RemoveContainer" containerID="0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7" Oct 10 16:24:10 crc kubenswrapper[4788]: E1010 16:24:10.980428 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7\": container with ID starting with 0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7 not found: ID does not exist" containerID="0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7" Oct 10 16:24:10 crc kubenswrapper[4788]: I1010 16:24:10.980459 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7"} err="failed to get container status \"0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7\": rpc error: code = NotFound desc = could not find container \"0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7\": container with ID starting with 0bc23ee97bd05a8eb4949253e694f7e1efb55f881277a00fa4eb1e74958afab7 not found: ID does not exist" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.206761 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.249702 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" path="/var/lib/kubelet/pods/d277f452-3e5f-4cb1-b511-bb79b496fc10/volumes" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.304715 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-7ffd598478-ntz5f"] Oct 10 16:24:12 crc kubenswrapper[4788]: E1010 16:24:12.305289 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerName="extract-utilities" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.305307 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerName="extract-utilities" Oct 10 16:24:12 crc kubenswrapper[4788]: E1010 16:24:12.305316 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a34bec5-c302-4e14-b422-669bba4d2ffb" containerName="ovn-config" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.305322 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a34bec5-c302-4e14-b422-669bba4d2ffb" containerName="ovn-config" Oct 10 16:24:12 crc kubenswrapper[4788]: E1010 16:24:12.305340 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8c72b5-9106-40ec-9b4e-6d993a0799c5" containerName="mariadb-account-create" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.305349 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8c72b5-9106-40ec-9b4e-6d993a0799c5" containerName="mariadb-account-create" Oct 10 16:24:12 crc kubenswrapper[4788]: E1010 16:24:12.305362 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerName="registry-server" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.305369 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerName="registry-server" Oct 10 16:24:12 crc kubenswrapper[4788]: E1010 16:24:12.305390 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerName="extract-content" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.305396 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerName="extract-content" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.305575 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8c72b5-9106-40ec-9b4e-6d993a0799c5" containerName="mariadb-account-create" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.305597 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a34bec5-c302-4e14-b422-669bba4d2ffb" containerName="ovn-config" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.305612 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d277f452-3e5f-4cb1-b511-bb79b496fc10" containerName="registry-server" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.316061 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.320423 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-px6mq" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.320661 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.320864 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.323589 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-7ffd598478-ntz5f"] Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.341787 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-log-ovn\") pod \"6a34bec5-c302-4e14-b422-669bba4d2ffb\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.341858 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run-ovn\") pod \"6a34bec5-c302-4e14-b422-669bba4d2ffb\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.341912 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6a34bec5-c302-4e14-b422-669bba4d2ffb" (UID: "6a34bec5-c302-4e14-b422-669bba4d2ffb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.341934 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-scripts\") pod \"6a34bec5-c302-4e14-b422-669bba4d2ffb\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.341956 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run\") pod \"6a34bec5-c302-4e14-b422-669bba4d2ffb\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.341991 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-additional-scripts\") pod \"6a34bec5-c302-4e14-b422-669bba4d2ffb\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.341987 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6a34bec5-c302-4e14-b422-669bba4d2ffb" (UID: "6a34bec5-c302-4e14-b422-669bba4d2ffb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.342022 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf9t6\" (UniqueName: \"kubernetes.io/projected/6a34bec5-c302-4e14-b422-669bba4d2ffb-kube-api-access-tf9t6\") pod \"6a34bec5-c302-4e14-b422-669bba4d2ffb\" (UID: \"6a34bec5-c302-4e14-b422-669bba4d2ffb\") " Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.342026 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run" (OuterVolumeSpecName: "var-run") pod "6a34bec5-c302-4e14-b422-669bba4d2ffb" (UID: "6a34bec5-c302-4e14-b422-669bba4d2ffb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.342537 4788 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.342553 4788 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.342563 4788 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a34bec5-c302-4e14-b422-669bba4d2ffb-var-run\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.342694 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6a34bec5-c302-4e14-b422-669bba4d2ffb" (UID: "6a34bec5-c302-4e14-b422-669bba4d2ffb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.343917 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-scripts" (OuterVolumeSpecName: "scripts") pod "6a34bec5-c302-4e14-b422-669bba4d2ffb" (UID: "6a34bec5-c302-4e14-b422-669bba4d2ffb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.363259 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a34bec5-c302-4e14-b422-669bba4d2ffb-kube-api-access-tf9t6" (OuterVolumeSpecName: "kube-api-access-tf9t6") pod "6a34bec5-c302-4e14-b422-669bba4d2ffb" (UID: "6a34bec5-c302-4e14-b422-669bba4d2ffb"). InnerVolumeSpecName "kube-api-access-tf9t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.443652 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/18a51038-256b-4638-9ca3-44a8c87718ab-config-data-merged\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.443698 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/18a51038-256b-4638-9ca3-44a8c87718ab-octavia-run\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.443859 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-scripts\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.444107 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-config-data\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.444384 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-combined-ca-bundle\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.444510 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.444527 4788 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a34bec5-c302-4e14-b422-669bba4d2ffb-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.444537 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf9t6\" (UniqueName: \"kubernetes.io/projected/6a34bec5-c302-4e14-b422-669bba4d2ffb-kube-api-access-tf9t6\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.546802 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-combined-ca-bundle\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.546849 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/18a51038-256b-4638-9ca3-44a8c87718ab-config-data-merged\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.546876 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/18a51038-256b-4638-9ca3-44a8c87718ab-octavia-run\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.546942 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-scripts\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.547441 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-config-data\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.547551 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/18a51038-256b-4638-9ca3-44a8c87718ab-config-data-merged\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.547603 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/18a51038-256b-4638-9ca3-44a8c87718ab-octavia-run\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.550047 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-scripts\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.550925 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-config-data\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.551128 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a51038-256b-4638-9ca3-44a8c87718ab-combined-ca-bundle\") pod \"octavia-api-7ffd598478-ntz5f\" (UID: \"18a51038-256b-4638-9ca3-44a8c87718ab\") " pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.642742 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.851226 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5k2fn-config-w7ljc" event={"ID":"6a34bec5-c302-4e14-b422-669bba4d2ffb","Type":"ContainerDied","Data":"a24a3bc2dab3969986f5fe648d671c4897a69e892f5777d42356ae4bd98d71bd"} Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.851727 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a24a3bc2dab3969986f5fe648d671c4897a69e892f5777d42356ae4bd98d71bd" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.851826 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5k2fn-config-w7ljc" Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.916899 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5k2fn-config-w7ljc"] Oct 10 16:24:12 crc kubenswrapper[4788]: I1010 16:24:12.942545 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5k2fn-config-w7ljc"] Oct 10 16:24:13 crc kubenswrapper[4788]: I1010 16:24:13.131237 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-7ffd598478-ntz5f"] Oct 10 16:24:13 crc kubenswrapper[4788]: W1010 16:24:13.133507 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18a51038_256b_4638_9ca3_44a8c87718ab.slice/crio-786587a85e3ec2ff34d880acb5f5eb6007fc29712fb1ebfde014c7b3f590e331 WatchSource:0}: Error finding container 786587a85e3ec2ff34d880acb5f5eb6007fc29712fb1ebfde014c7b3f590e331: Status 404 returned error can't find the container with id 786587a85e3ec2ff34d880acb5f5eb6007fc29712fb1ebfde014c7b3f590e331 Oct 10 16:24:13 crc kubenswrapper[4788]: I1010 16:24:13.325576 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-5k2fn" Oct 10 16:24:13 crc kubenswrapper[4788]: I1010 16:24:13.860112 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7ffd598478-ntz5f" event={"ID":"18a51038-256b-4638-9ca3-44a8c87718ab","Type":"ContainerStarted","Data":"786587a85e3ec2ff34d880acb5f5eb6007fc29712fb1ebfde014c7b3f590e331"} Oct 10 16:24:14 crc kubenswrapper[4788]: I1010 16:24:14.248327 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a34bec5-c302-4e14-b422-669bba4d2ffb" path="/var/lib/kubelet/pods/6a34bec5-c302-4e14-b422-669bba4d2ffb/volumes" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.698240 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f275r"] Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.700865 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.710774 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f275r"] Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.790801 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-utilities\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.790868 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-catalog-content\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.790938 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7vtq\" (UniqueName: \"kubernetes.io/projected/29253b54-6691-4169-a6f6-b6b0fd6e9d35-kube-api-access-z7vtq\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.892747 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7vtq\" (UniqueName: \"kubernetes.io/projected/29253b54-6691-4169-a6f6-b6b0fd6e9d35-kube-api-access-z7vtq\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.892897 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-utilities\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.892936 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-catalog-content\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.893493 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-utilities\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.893561 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-catalog-content\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:24 crc kubenswrapper[4788]: I1010 16:24:24.918096 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7vtq\" (UniqueName: \"kubernetes.io/projected/29253b54-6691-4169-a6f6-b6b0fd6e9d35-kube-api-access-z7vtq\") pod \"redhat-marketplace-f275r\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:25 crc kubenswrapper[4788]: I1010 16:24:25.024424 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:28 crc kubenswrapper[4788]: I1010 16:24:28.032660 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f275r"] Oct 10 16:24:28 crc kubenswrapper[4788]: I1010 16:24:28.035283 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7ffd598478-ntz5f" event={"ID":"18a51038-256b-4638-9ca3-44a8c87718ab","Type":"ContainerStarted","Data":"ba5329cc4b5e0b36e9d69bcf46137896b533828120338a41a8c78d563d99d5f2"} Oct 10 16:24:28 crc kubenswrapper[4788]: W1010 16:24:28.036831 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29253b54_6691_4169_a6f6_b6b0fd6e9d35.slice/crio-25dae08d55f210829c8bb21bda60f0d0a8c6ca27a4f08620c31dc9ecbb75ea2b WatchSource:0}: Error finding container 25dae08d55f210829c8bb21bda60f0d0a8c6ca27a4f08620c31dc9ecbb75ea2b: Status 404 returned error can't find the container with id 25dae08d55f210829c8bb21bda60f0d0a8c6ca27a4f08620c31dc9ecbb75ea2b Oct 10 16:24:29 crc kubenswrapper[4788]: I1010 16:24:29.045693 4788 generic.go:334] "Generic (PLEG): container finished" podID="18a51038-256b-4638-9ca3-44a8c87718ab" containerID="ba5329cc4b5e0b36e9d69bcf46137896b533828120338a41a8c78d563d99d5f2" exitCode=0 Oct 10 16:24:29 crc kubenswrapper[4788]: I1010 16:24:29.045772 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7ffd598478-ntz5f" event={"ID":"18a51038-256b-4638-9ca3-44a8c87718ab","Type":"ContainerDied","Data":"ba5329cc4b5e0b36e9d69bcf46137896b533828120338a41a8c78d563d99d5f2"} Oct 10 16:24:29 crc kubenswrapper[4788]: I1010 16:24:29.050693 4788 generic.go:334] "Generic (PLEG): container finished" podID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerID="d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41" exitCode=0 Oct 10 16:24:29 crc kubenswrapper[4788]: I1010 16:24:29.050782 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f275r" event={"ID":"29253b54-6691-4169-a6f6-b6b0fd6e9d35","Type":"ContainerDied","Data":"d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41"} Oct 10 16:24:29 crc kubenswrapper[4788]: I1010 16:24:29.050836 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f275r" event={"ID":"29253b54-6691-4169-a6f6-b6b0fd6e9d35","Type":"ContainerStarted","Data":"25dae08d55f210829c8bb21bda60f0d0a8c6ca27a4f08620c31dc9ecbb75ea2b"} Oct 10 16:24:30 crc kubenswrapper[4788]: I1010 16:24:30.061732 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f275r" event={"ID":"29253b54-6691-4169-a6f6-b6b0fd6e9d35","Type":"ContainerStarted","Data":"674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887"} Oct 10 16:24:30 crc kubenswrapper[4788]: I1010 16:24:30.070075 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7ffd598478-ntz5f" event={"ID":"18a51038-256b-4638-9ca3-44a8c87718ab","Type":"ContainerStarted","Data":"4af184daa7e3e538c82389a1503719c65ef266e98e99e30f16c34ea0ae9c49a2"} Oct 10 16:24:30 crc kubenswrapper[4788]: I1010 16:24:30.070255 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7ffd598478-ntz5f" event={"ID":"18a51038-256b-4638-9ca3-44a8c87718ab","Type":"ContainerStarted","Data":"1ea0a49f697b1b7ceadf8da09e13ad20e0f34318e71d9394949d7b3f00530c2b"} Oct 10 16:24:30 crc kubenswrapper[4788]: I1010 16:24:30.070468 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:30 crc kubenswrapper[4788]: I1010 16:24:30.139394 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-7ffd598478-ntz5f" podStartSLOduration=3.6320470719999998 podStartE2EDuration="18.139375681s" podCreationTimestamp="2025-10-10 16:24:12 +0000 UTC" firstStartedPulling="2025-10-10 16:24:13.136213194 +0000 UTC m=+5955.585928742" lastFinishedPulling="2025-10-10 16:24:27.643541793 +0000 UTC m=+5970.093257351" observedRunningTime="2025-10-10 16:24:30.129344483 +0000 UTC m=+5972.579060031" watchObservedRunningTime="2025-10-10 16:24:30.139375681 +0000 UTC m=+5972.589091229" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.087597 4788 generic.go:334] "Generic (PLEG): container finished" podID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerID="674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887" exitCode=0 Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.089063 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f275r" event={"ID":"29253b54-6691-4169-a6f6-b6b0fd6e9d35","Type":"ContainerDied","Data":"674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887"} Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.089642 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.780105 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-jttq6"] Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.782342 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.784384 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.784761 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.788708 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.807489 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-jttq6"] Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.856261 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9efa8985-a060-40c9-ba49-47e7d866c630-config-data-merged\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.856311 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efa8985-a060-40c9-ba49-47e7d866c630-config-data\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.856365 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efa8985-a060-40c9-ba49-47e7d866c630-scripts\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.856392 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/9efa8985-a060-40c9-ba49-47e7d866c630-hm-ports\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.959329 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efa8985-a060-40c9-ba49-47e7d866c630-scripts\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.959585 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/9efa8985-a060-40c9-ba49-47e7d866c630-hm-ports\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.959853 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9efa8985-a060-40c9-ba49-47e7d866c630-config-data-merged\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.959973 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efa8985-a060-40c9-ba49-47e7d866c630-config-data\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.960572 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/9efa8985-a060-40c9-ba49-47e7d866c630-config-data-merged\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.961238 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/9efa8985-a060-40c9-ba49-47e7d866c630-hm-ports\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.966342 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efa8985-a060-40c9-ba49-47e7d866c630-config-data\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:31 crc kubenswrapper[4788]: I1010 16:24:31.986313 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efa8985-a060-40c9-ba49-47e7d866c630-scripts\") pod \"octavia-rsyslog-jttq6\" (UID: \"9efa8985-a060-40c9-ba49-47e7d866c630\") " pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.104683 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f275r" event={"ID":"29253b54-6691-4169-a6f6-b6b0fd6e9d35","Type":"ContainerStarted","Data":"6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664"} Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.125075 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f275r" podStartSLOduration=5.463402475 podStartE2EDuration="8.125055792s" podCreationTimestamp="2025-10-10 16:24:24 +0000 UTC" firstStartedPulling="2025-10-10 16:24:29.05373487 +0000 UTC m=+5971.503450458" lastFinishedPulling="2025-10-10 16:24:31.715388227 +0000 UTC m=+5974.165103775" observedRunningTime="2025-10-10 16:24:32.124262661 +0000 UTC m=+5974.573978219" watchObservedRunningTime="2025-10-10 16:24:32.125055792 +0000 UTC m=+5974.574771340" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.191128 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.732240 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-9fpkw"] Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.734579 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.737916 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.743372 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-9fpkw"] Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.824003 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-jttq6"] Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.875351 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a501692c-d1f3-4ace-abf8-a73d2ad83102-amphora-image\") pod \"octavia-image-upload-59f8cff499-9fpkw\" (UID: \"a501692c-d1f3-4ace-abf8-a73d2ad83102\") " pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.875742 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a501692c-d1f3-4ace-abf8-a73d2ad83102-httpd-config\") pod \"octavia-image-upload-59f8cff499-9fpkw\" (UID: \"a501692c-d1f3-4ace-abf8-a73d2ad83102\") " pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.977252 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a501692c-d1f3-4ace-abf8-a73d2ad83102-httpd-config\") pod \"octavia-image-upload-59f8cff499-9fpkw\" (UID: \"a501692c-d1f3-4ace-abf8-a73d2ad83102\") " pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.977354 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a501692c-d1f3-4ace-abf8-a73d2ad83102-amphora-image\") pod \"octavia-image-upload-59f8cff499-9fpkw\" (UID: \"a501692c-d1f3-4ace-abf8-a73d2ad83102\") " pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.977900 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a501692c-d1f3-4ace-abf8-a73d2ad83102-amphora-image\") pod \"octavia-image-upload-59f8cff499-9fpkw\" (UID: \"a501692c-d1f3-4ace-abf8-a73d2ad83102\") " pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:24:32 crc kubenswrapper[4788]: I1010 16:24:32.982955 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a501692c-d1f3-4ace-abf8-a73d2ad83102-httpd-config\") pod \"octavia-image-upload-59f8cff499-9fpkw\" (UID: \"a501692c-d1f3-4ace-abf8-a73d2ad83102\") " pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:24:33 crc kubenswrapper[4788]: I1010 16:24:33.058044 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:24:33 crc kubenswrapper[4788]: I1010 16:24:33.117979 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-jttq6" event={"ID":"9efa8985-a060-40c9-ba49-47e7d866c630","Type":"ContainerStarted","Data":"ec7c90ccf077d802b85a10b215ce695584fb9b3f23b01b7beb451327b41ee408"} Oct 10 16:24:33 crc kubenswrapper[4788]: I1010 16:24:33.497427 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-9fpkw"] Oct 10 16:24:34 crc kubenswrapper[4788]: I1010 16:24:34.129097 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" event={"ID":"a501692c-d1f3-4ace-abf8-a73d2ad83102","Type":"ContainerStarted","Data":"61eb80b3802b0e3208867614314d11b6cc34f00eb26914905bc3a38e881dd831"} Oct 10 16:24:35 crc kubenswrapper[4788]: I1010 16:24:35.024656 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:35 crc kubenswrapper[4788]: I1010 16:24:35.025400 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:35 crc kubenswrapper[4788]: I1010 16:24:35.085292 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:39 crc kubenswrapper[4788]: I1010 16:24:39.209179 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-jttq6" event={"ID":"9efa8985-a060-40c9-ba49-47e7d866c630","Type":"ContainerStarted","Data":"b780dcbc4689308b2cc7ada41bf6ca496a555c7ac589d7a311170f9845e2e4cb"} Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.101032 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-582jp"] Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.103891 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.107601 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.124779 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-582jp"] Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.275391 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-config-data\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.275527 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-combined-ca-bundle\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.275559 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-scripts\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.275583 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/18752d95-0174-4746-8554-00be4abbcaa7-config-data-merged\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.377574 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-config-data\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.377817 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-combined-ca-bundle\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.377859 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-scripts\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.377889 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/18752d95-0174-4746-8554-00be4abbcaa7-config-data-merged\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.379045 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/18752d95-0174-4746-8554-00be4abbcaa7-config-data-merged\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.385297 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-scripts\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.386208 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-config-data\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.400477 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-combined-ca-bundle\") pod \"octavia-db-sync-582jp\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.446018 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:40 crc kubenswrapper[4788]: I1010 16:24:40.943598 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-582jp"] Oct 10 16:24:42 crc kubenswrapper[4788]: I1010 16:24:42.245512 4788 generic.go:334] "Generic (PLEG): container finished" podID="9efa8985-a060-40c9-ba49-47e7d866c630" containerID="b780dcbc4689308b2cc7ada41bf6ca496a555c7ac589d7a311170f9845e2e4cb" exitCode=0 Oct 10 16:24:42 crc kubenswrapper[4788]: I1010 16:24:42.245782 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-jttq6" event={"ID":"9efa8985-a060-40c9-ba49-47e7d866c630","Type":"ContainerDied","Data":"b780dcbc4689308b2cc7ada41bf6ca496a555c7ac589d7a311170f9845e2e4cb"} Oct 10 16:24:45 crc kubenswrapper[4788]: I1010 16:24:45.074902 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:45 crc kubenswrapper[4788]: I1010 16:24:45.122803 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f275r"] Oct 10 16:24:45 crc kubenswrapper[4788]: I1010 16:24:45.282508 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f275r" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerName="registry-server" containerID="cri-o://6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664" gracePeriod=2 Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.099014 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.206883 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7vtq\" (UniqueName: \"kubernetes.io/projected/29253b54-6691-4169-a6f6-b6b0fd6e9d35-kube-api-access-z7vtq\") pod \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.207294 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-catalog-content\") pod \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.207573 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-utilities\") pod \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\" (UID: \"29253b54-6691-4169-a6f6-b6b0fd6e9d35\") " Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.208015 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-utilities" (OuterVolumeSpecName: "utilities") pod "29253b54-6691-4169-a6f6-b6b0fd6e9d35" (UID: "29253b54-6691-4169-a6f6-b6b0fd6e9d35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.208514 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.212899 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29253b54-6691-4169-a6f6-b6b0fd6e9d35-kube-api-access-z7vtq" (OuterVolumeSpecName: "kube-api-access-z7vtq") pod "29253b54-6691-4169-a6f6-b6b0fd6e9d35" (UID: "29253b54-6691-4169-a6f6-b6b0fd6e9d35"). InnerVolumeSpecName "kube-api-access-z7vtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.219225 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29253b54-6691-4169-a6f6-b6b0fd6e9d35" (UID: "29253b54-6691-4169-a6f6-b6b0fd6e9d35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.310089 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-582jp" event={"ID":"18752d95-0174-4746-8554-00be4abbcaa7","Type":"ContainerStarted","Data":"d05d0681a13d9545bb1158106066a0fae930cb485a42a6d73ce005e91ee8cee7"} Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.310329 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-582jp" event={"ID":"18752d95-0174-4746-8554-00be4abbcaa7","Type":"ContainerStarted","Data":"658b589e19a9c45e83ddbda9b58ca6cee27f534d11a26cc38fc00cf1012d784a"} Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.311809 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29253b54-6691-4169-a6f6-b6b0fd6e9d35-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.311843 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7vtq\" (UniqueName: \"kubernetes.io/projected/29253b54-6691-4169-a6f6-b6b0fd6e9d35-kube-api-access-z7vtq\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.317541 4788 generic.go:334] "Generic (PLEG): container finished" podID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerID="6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664" exitCode=0 Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.317697 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f275r" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.318635 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f275r" event={"ID":"29253b54-6691-4169-a6f6-b6b0fd6e9d35","Type":"ContainerDied","Data":"6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664"} Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.318708 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f275r" event={"ID":"29253b54-6691-4169-a6f6-b6b0fd6e9d35","Type":"ContainerDied","Data":"25dae08d55f210829c8bb21bda60f0d0a8c6ca27a4f08620c31dc9ecbb75ea2b"} Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.318730 4788 scope.go:117] "RemoveContainer" containerID="6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.327938 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" event={"ID":"a501692c-d1f3-4ace-abf8-a73d2ad83102","Type":"ContainerStarted","Data":"2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a"} Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.354927 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f275r"] Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.355356 4788 scope.go:117] "RemoveContainer" containerID="674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.366953 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f275r"] Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.386916 4788 scope.go:117] "RemoveContainer" containerID="d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.462850 4788 scope.go:117] "RemoveContainer" containerID="6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664" Oct 10 16:24:46 crc kubenswrapper[4788]: E1010 16:24:46.463359 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664\": container with ID starting with 6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664 not found: ID does not exist" containerID="6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.463393 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664"} err="failed to get container status \"6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664\": rpc error: code = NotFound desc = could not find container \"6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664\": container with ID starting with 6e31cc6e365733bc016ecc12b308d0cc73342222d98a919dbdd7bf98a6742664 not found: ID does not exist" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.463416 4788 scope.go:117] "RemoveContainer" containerID="674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887" Oct 10 16:24:46 crc kubenswrapper[4788]: E1010 16:24:46.463718 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887\": container with ID starting with 674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887 not found: ID does not exist" containerID="674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.463740 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887"} err="failed to get container status \"674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887\": rpc error: code = NotFound desc = could not find container \"674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887\": container with ID starting with 674a274ca12e5bb5b6e663cba428bef06a8b834255d06e5681d4dcb33ef0e887 not found: ID does not exist" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.463754 4788 scope.go:117] "RemoveContainer" containerID="d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41" Oct 10 16:24:46 crc kubenswrapper[4788]: E1010 16:24:46.464075 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41\": container with ID starting with d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41 not found: ID does not exist" containerID="d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.464119 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41"} err="failed to get container status \"d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41\": rpc error: code = NotFound desc = could not find container \"d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41\": container with ID starting with d29d37194aa320fcca245e40d76d95140b0af909aa073d17315e40a1edff4e41 not found: ID does not exist" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.677523 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:46 crc kubenswrapper[4788]: I1010 16:24:46.719644 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-7ffd598478-ntz5f" Oct 10 16:24:47 crc kubenswrapper[4788]: I1010 16:24:47.338594 4788 generic.go:334] "Generic (PLEG): container finished" podID="18752d95-0174-4746-8554-00be4abbcaa7" containerID="d05d0681a13d9545bb1158106066a0fae930cb485a42a6d73ce005e91ee8cee7" exitCode=0 Oct 10 16:24:47 crc kubenswrapper[4788]: I1010 16:24:47.338703 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-582jp" event={"ID":"18752d95-0174-4746-8554-00be4abbcaa7","Type":"ContainerDied","Data":"d05d0681a13d9545bb1158106066a0fae930cb485a42a6d73ce005e91ee8cee7"} Oct 10 16:24:48 crc kubenswrapper[4788]: I1010 16:24:48.259252 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" path="/var/lib/kubelet/pods/29253b54-6691-4169-a6f6-b6b0fd6e9d35/volumes" Oct 10 16:24:48 crc kubenswrapper[4788]: I1010 16:24:48.365374 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-jttq6" event={"ID":"9efa8985-a060-40c9-ba49-47e7d866c630","Type":"ContainerStarted","Data":"2bf844f973163026a21acacec7ee1f0ed75a4e57f4a1ee266f088d11466f55c7"} Oct 10 16:24:49 crc kubenswrapper[4788]: I1010 16:24:49.376170 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-582jp" event={"ID":"18752d95-0174-4746-8554-00be4abbcaa7","Type":"ContainerStarted","Data":"dec256bd3c52330cd0cce3600810ca50b9258497051c772792dc53bff210c1d9"} Oct 10 16:24:49 crc kubenswrapper[4788]: I1010 16:24:49.377446 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:24:49 crc kubenswrapper[4788]: I1010 16:24:49.399413 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-582jp" podStartSLOduration=9.399392885 podStartE2EDuration="9.399392885s" podCreationTimestamp="2025-10-10 16:24:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:24:49.394727196 +0000 UTC m=+5991.844442744" watchObservedRunningTime="2025-10-10 16:24:49.399392885 +0000 UTC m=+5991.849108473" Oct 10 16:24:49 crc kubenswrapper[4788]: I1010 16:24:49.413570 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-jttq6" podStartSLOduration=4.14640912 podStartE2EDuration="18.41355202s" podCreationTimestamp="2025-10-10 16:24:31 +0000 UTC" firstStartedPulling="2025-10-10 16:24:32.826303347 +0000 UTC m=+5975.276018895" lastFinishedPulling="2025-10-10 16:24:47.093446247 +0000 UTC m=+5989.543161795" observedRunningTime="2025-10-10 16:24:49.410991954 +0000 UTC m=+5991.860707542" watchObservedRunningTime="2025-10-10 16:24:49.41355202 +0000 UTC m=+5991.863267608" Oct 10 16:24:51 crc kubenswrapper[4788]: I1010 16:24:51.398975 4788 generic.go:334] "Generic (PLEG): container finished" podID="a501692c-d1f3-4ace-abf8-a73d2ad83102" containerID="2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a" exitCode=0 Oct 10 16:24:51 crc kubenswrapper[4788]: I1010 16:24:51.399057 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" event={"ID":"a501692c-d1f3-4ace-abf8-a73d2ad83102","Type":"ContainerDied","Data":"2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a"} Oct 10 16:24:52 crc kubenswrapper[4788]: I1010 16:24:52.411523 4788 generic.go:334] "Generic (PLEG): container finished" podID="18752d95-0174-4746-8554-00be4abbcaa7" containerID="dec256bd3c52330cd0cce3600810ca50b9258497051c772792dc53bff210c1d9" exitCode=0 Oct 10 16:24:52 crc kubenswrapper[4788]: I1010 16:24:52.411615 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-582jp" event={"ID":"18752d95-0174-4746-8554-00be4abbcaa7","Type":"ContainerDied","Data":"dec256bd3c52330cd0cce3600810ca50b9258497051c772792dc53bff210c1d9"} Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.423920 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" event={"ID":"a501692c-d1f3-4ace-abf8-a73d2ad83102","Type":"ContainerStarted","Data":"fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843"} Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.445690 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" podStartSLOduration=2.063352866 podStartE2EDuration="21.445666073s" podCreationTimestamp="2025-10-10 16:24:32 +0000 UTC" firstStartedPulling="2025-10-10 16:24:33.510936215 +0000 UTC m=+5975.960651753" lastFinishedPulling="2025-10-10 16:24:52.893249372 +0000 UTC m=+5995.342964960" observedRunningTime="2025-10-10 16:24:53.438861059 +0000 UTC m=+5995.888576607" watchObservedRunningTime="2025-10-10 16:24:53.445666073 +0000 UTC m=+5995.895381621" Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.809849 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-582jp" Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.967690 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-combined-ca-bundle\") pod \"18752d95-0174-4746-8554-00be4abbcaa7\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.968129 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/18752d95-0174-4746-8554-00be4abbcaa7-config-data-merged\") pod \"18752d95-0174-4746-8554-00be4abbcaa7\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.968259 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-config-data\") pod \"18752d95-0174-4746-8554-00be4abbcaa7\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.968439 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-scripts\") pod \"18752d95-0174-4746-8554-00be4abbcaa7\" (UID: \"18752d95-0174-4746-8554-00be4abbcaa7\") " Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.982448 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-config-data" (OuterVolumeSpecName: "config-data") pod "18752d95-0174-4746-8554-00be4abbcaa7" (UID: "18752d95-0174-4746-8554-00be4abbcaa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.982541 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-scripts" (OuterVolumeSpecName: "scripts") pod "18752d95-0174-4746-8554-00be4abbcaa7" (UID: "18752d95-0174-4746-8554-00be4abbcaa7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:24:53 crc kubenswrapper[4788]: I1010 16:24:53.998856 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18752d95-0174-4746-8554-00be4abbcaa7-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "18752d95-0174-4746-8554-00be4abbcaa7" (UID: "18752d95-0174-4746-8554-00be4abbcaa7"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:24:54 crc kubenswrapper[4788]: I1010 16:24:54.007689 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18752d95-0174-4746-8554-00be4abbcaa7" (UID: "18752d95-0174-4746-8554-00be4abbcaa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:24:54 crc kubenswrapper[4788]: I1010 16:24:54.070637 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:54 crc kubenswrapper[4788]: I1010 16:24:54.070670 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:54 crc kubenswrapper[4788]: I1010 16:24:54.070679 4788 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/18752d95-0174-4746-8554-00be4abbcaa7-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:54 crc kubenswrapper[4788]: I1010 16:24:54.070687 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18752d95-0174-4746-8554-00be4abbcaa7-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:24:54 crc kubenswrapper[4788]: I1010 16:24:54.434016 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-582jp" event={"ID":"18752d95-0174-4746-8554-00be4abbcaa7","Type":"ContainerDied","Data":"658b589e19a9c45e83ddbda9b58ca6cee27f534d11a26cc38fc00cf1012d784a"} Oct 10 16:24:54 crc kubenswrapper[4788]: I1010 16:24:54.434071 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="658b589e19a9c45e83ddbda9b58ca6cee27f534d11a26cc38fc00cf1012d784a" Oct 10 16:24:54 crc kubenswrapper[4788]: I1010 16:24:54.434187 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-582jp" Oct 10 16:25:02 crc kubenswrapper[4788]: I1010 16:25:02.248894 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-jttq6" Oct 10 16:25:09 crc kubenswrapper[4788]: I1010 16:25:09.052025 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-cmgzc"] Oct 10 16:25:09 crc kubenswrapper[4788]: I1010 16:25:09.064873 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-cmgzc"] Oct 10 16:25:10 crc kubenswrapper[4788]: I1010 16:25:10.244366 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6890f9fa-fe41-4306-b929-333908049b5b" path="/var/lib/kubelet/pods/6890f9fa-fe41-4306-b929-333908049b5b/volumes" Oct 10 16:25:13 crc kubenswrapper[4788]: I1010 16:25:13.740563 4788 scope.go:117] "RemoveContainer" containerID="0a9db373359db6e784c743683b81485d53adfe54e92be33fec9f3c29bf31f207" Oct 10 16:25:19 crc kubenswrapper[4788]: I1010 16:25:19.041761 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-75a0-account-create-87fb5"] Oct 10 16:25:19 crc kubenswrapper[4788]: I1010 16:25:19.050613 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-75a0-account-create-87fb5"] Oct 10 16:25:20 crc kubenswrapper[4788]: I1010 16:25:20.245389 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9abf38c2-4b60-49d2-9170-1e94f8f85dcb" path="/var/lib/kubelet/pods/9abf38c2-4b60-49d2-9170-1e94f8f85dcb/volumes" Oct 10 16:25:20 crc kubenswrapper[4788]: I1010 16:25:20.455126 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-9fpkw"] Oct 10 16:25:20 crc kubenswrapper[4788]: I1010 16:25:20.455407 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" podUID="a501692c-d1f3-4ace-abf8-a73d2ad83102" containerName="octavia-amphora-httpd" containerID="cri-o://fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843" gracePeriod=30 Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.340592 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.474401 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a501692c-d1f3-4ace-abf8-a73d2ad83102-amphora-image\") pod \"a501692c-d1f3-4ace-abf8-a73d2ad83102\" (UID: \"a501692c-d1f3-4ace-abf8-a73d2ad83102\") " Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.474559 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a501692c-d1f3-4ace-abf8-a73d2ad83102-httpd-config\") pod \"a501692c-d1f3-4ace-abf8-a73d2ad83102\" (UID: \"a501692c-d1f3-4ace-abf8-a73d2ad83102\") " Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.499434 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a501692c-d1f3-4ace-abf8-a73d2ad83102-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "a501692c-d1f3-4ace-abf8-a73d2ad83102" (UID: "a501692c-d1f3-4ace-abf8-a73d2ad83102"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.501059 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a501692c-d1f3-4ace-abf8-a73d2ad83102-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a501692c-d1f3-4ace-abf8-a73d2ad83102" (UID: "a501692c-d1f3-4ace-abf8-a73d2ad83102"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.577794 4788 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a501692c-d1f3-4ace-abf8-a73d2ad83102-amphora-image\") on node \"crc\" DevicePath \"\"" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.577839 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a501692c-d1f3-4ace-abf8-a73d2ad83102-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.734891 4788 generic.go:334] "Generic (PLEG): container finished" podID="a501692c-d1f3-4ace-abf8-a73d2ad83102" containerID="fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843" exitCode=0 Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.734969 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.734963 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" event={"ID":"a501692c-d1f3-4ace-abf8-a73d2ad83102","Type":"ContainerDied","Data":"fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843"} Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.736413 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-9fpkw" event={"ID":"a501692c-d1f3-4ace-abf8-a73d2ad83102","Type":"ContainerDied","Data":"61eb80b3802b0e3208867614314d11b6cc34f00eb26914905bc3a38e881dd831"} Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.736436 4788 scope.go:117] "RemoveContainer" containerID="fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.767640 4788 scope.go:117] "RemoveContainer" containerID="2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.772339 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-9fpkw"] Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.792601 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-9fpkw"] Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.801271 4788 scope.go:117] "RemoveContainer" containerID="fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843" Oct 10 16:25:21 crc kubenswrapper[4788]: E1010 16:25:21.801772 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843\": container with ID starting with fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843 not found: ID does not exist" containerID="fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.801816 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843"} err="failed to get container status \"fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843\": rpc error: code = NotFound desc = could not find container \"fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843\": container with ID starting with fb38df014385fec46bd75d3a565faa3c3a70aa9532a48aa255eef55fa2430843 not found: ID does not exist" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.801840 4788 scope.go:117] "RemoveContainer" containerID="2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a" Oct 10 16:25:21 crc kubenswrapper[4788]: E1010 16:25:21.802354 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a\": container with ID starting with 2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a not found: ID does not exist" containerID="2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a" Oct 10 16:25:21 crc kubenswrapper[4788]: I1010 16:25:21.802390 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a"} err="failed to get container status \"2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a\": rpc error: code = NotFound desc = could not find container \"2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a\": container with ID starting with 2681ff5ceed49fde87084bed59dc17d6c6dbe16cb60c26647d34f663d536f86a not found: ID does not exist" Oct 10 16:25:22 crc kubenswrapper[4788]: I1010 16:25:22.243589 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a501692c-d1f3-4ace-abf8-a73d2ad83102" path="/var/lib/kubelet/pods/a501692c-d1f3-4ace-abf8-a73d2ad83102/volumes" Oct 10 16:25:25 crc kubenswrapper[4788]: I1010 16:25:25.032431 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-z6wgq"] Oct 10 16:25:25 crc kubenswrapper[4788]: I1010 16:25:25.042817 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-z6wgq"] Oct 10 16:25:26 crc kubenswrapper[4788]: I1010 16:25:26.250995 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2189f91-f8ed-47ef-8b6e-00b20b6b996b" path="/var/lib/kubelet/pods/e2189f91-f8ed-47ef-8b6e-00b20b6b996b/volumes" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.820629 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-dlvvq"] Oct 10 16:25:38 crc kubenswrapper[4788]: E1010 16:25:38.822545 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18752d95-0174-4746-8554-00be4abbcaa7" containerName="init" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.822573 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="18752d95-0174-4746-8554-00be4abbcaa7" containerName="init" Oct 10 16:25:38 crc kubenswrapper[4788]: E1010 16:25:38.822615 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerName="extract-content" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.822628 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerName="extract-content" Oct 10 16:25:38 crc kubenswrapper[4788]: E1010 16:25:38.822659 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerName="registry-server" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.822672 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerName="registry-server" Oct 10 16:25:38 crc kubenswrapper[4788]: E1010 16:25:38.822698 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a501692c-d1f3-4ace-abf8-a73d2ad83102" containerName="init" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.822710 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a501692c-d1f3-4ace-abf8-a73d2ad83102" containerName="init" Oct 10 16:25:38 crc kubenswrapper[4788]: E1010 16:25:38.822763 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18752d95-0174-4746-8554-00be4abbcaa7" containerName="octavia-db-sync" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.822776 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="18752d95-0174-4746-8554-00be4abbcaa7" containerName="octavia-db-sync" Oct 10 16:25:38 crc kubenswrapper[4788]: E1010 16:25:38.822799 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerName="extract-utilities" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.822812 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerName="extract-utilities" Oct 10 16:25:38 crc kubenswrapper[4788]: E1010 16:25:38.822827 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a501692c-d1f3-4ace-abf8-a73d2ad83102" containerName="octavia-amphora-httpd" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.822839 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a501692c-d1f3-4ace-abf8-a73d2ad83102" containerName="octavia-amphora-httpd" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.823279 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="18752d95-0174-4746-8554-00be4abbcaa7" containerName="octavia-db-sync" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.823303 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="29253b54-6691-4169-a6f6-b6b0fd6e9d35" containerName="registry-server" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.823317 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a501692c-d1f3-4ace-abf8-a73d2ad83102" containerName="octavia-amphora-httpd" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.825488 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.828791 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.828902 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.829737 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.843061 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-dlvvq"] Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.981621 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-amphora-certs\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.982070 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-config-data-merged\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.982104 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-scripts\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.982156 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-combined-ca-bundle\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.982360 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-config-data\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:38 crc kubenswrapper[4788]: I1010 16:25:38.982400 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-hm-ports\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.084265 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-config-data-merged\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.084371 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-scripts\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.084461 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-combined-ca-bundle\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.084637 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-config-data\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.085055 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-config-data-merged\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.085752 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-hm-ports\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.085958 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-amphora-certs\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.087052 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-hm-ports\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.092265 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-scripts\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.092725 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-amphora-certs\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.092800 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-combined-ca-bundle\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.102103 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1-config-data\") pod \"octavia-healthmanager-dlvvq\" (UID: \"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1\") " pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.166179 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.767937 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-dlvvq"] Oct 10 16:25:39 crc kubenswrapper[4788]: W1010 16:25:39.768828 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dd3f0aa_e9c2_4d5b_a2d2_f02df5fe23c1.slice/crio-ae5498a15a97977a8b71d866998e35d000ce5d23340665590f64f9d28cb5d549 WatchSource:0}: Error finding container ae5498a15a97977a8b71d866998e35d000ce5d23340665590f64f9d28cb5d549: Status 404 returned error can't find the container with id ae5498a15a97977a8b71d866998e35d000ce5d23340665590f64f9d28cb5d549 Oct 10 16:25:39 crc kubenswrapper[4788]: I1010 16:25:39.921854 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-dlvvq" event={"ID":"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1","Type":"ContainerStarted","Data":"ae5498a15a97977a8b71d866998e35d000ce5d23340665590f64f9d28cb5d549"} Oct 10 16:25:40 crc kubenswrapper[4788]: I1010 16:25:40.935432 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-dlvvq" event={"ID":"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1","Type":"ContainerStarted","Data":"807d53f0ec2463e769400d4af8c12ed79ba220086ec21e2a7b8ab8359adf8044"} Oct 10 16:25:40 crc kubenswrapper[4788]: I1010 16:25:40.954830 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-9n42v"] Oct 10 16:25:40 crc kubenswrapper[4788]: I1010 16:25:40.957308 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:40 crc kubenswrapper[4788]: I1010 16:25:40.963622 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Oct 10 16:25:40 crc kubenswrapper[4788]: I1010 16:25:40.969245 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Oct 10 16:25:40 crc kubenswrapper[4788]: I1010 16:25:40.978273 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-9n42v"] Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.061288 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e74a446b-ff90-447e-82c4-a61d927dc0a3-config-data-merged\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.062582 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-config-data\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.062652 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-amphora-certs\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.062783 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-scripts\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.062865 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-combined-ca-bundle\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.062897 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e74a446b-ff90-447e-82c4-a61d927dc0a3-hm-ports\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.164884 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-config-data\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.165551 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-amphora-certs\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.165723 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-scripts\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.165887 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-combined-ca-bundle\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.166002 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e74a446b-ff90-447e-82c4-a61d927dc0a3-hm-ports\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.166206 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e74a446b-ff90-447e-82c4-a61d927dc0a3-config-data-merged\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.166879 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e74a446b-ff90-447e-82c4-a61d927dc0a3-config-data-merged\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.168223 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/e74a446b-ff90-447e-82c4-a61d927dc0a3-hm-ports\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.171389 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-combined-ca-bundle\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.171510 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-config-data\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.172312 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-amphora-certs\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.185231 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74a446b-ff90-447e-82c4-a61d927dc0a3-scripts\") pod \"octavia-housekeeping-9n42v\" (UID: \"e74a446b-ff90-447e-82c4-a61d927dc0a3\") " pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.295880 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.676894 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-9n42v"] Oct 10 16:25:41 crc kubenswrapper[4788]: I1010 16:25:41.947192 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-9n42v" event={"ID":"e74a446b-ff90-447e-82c4-a61d927dc0a3","Type":"ContainerStarted","Data":"05bd34a836c4fcb96138f7393e844c03894fec707ce2ee9590c9cfb842866ca4"} Oct 10 16:25:42 crc kubenswrapper[4788]: I1010 16:25:42.959475 4788 generic.go:334] "Generic (PLEG): container finished" podID="6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1" containerID="807d53f0ec2463e769400d4af8c12ed79ba220086ec21e2a7b8ab8359adf8044" exitCode=0 Oct 10 16:25:42 crc kubenswrapper[4788]: I1010 16:25:42.959524 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-dlvvq" event={"ID":"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1","Type":"ContainerDied","Data":"807d53f0ec2463e769400d4af8c12ed79ba220086ec21e2a7b8ab8359adf8044"} Oct 10 16:25:43 crc kubenswrapper[4788]: I1010 16:25:43.979874 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-dlvvq" event={"ID":"6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1","Type":"ContainerStarted","Data":"387ed59db6b951dc1bfa89ae4467bd6fdb8206e113dd76446608ff0832ae5990"} Oct 10 16:25:43 crc kubenswrapper[4788]: I1010 16:25:43.980848 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.003658 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-dlvvq" podStartSLOduration=6.003637349 podStartE2EDuration="6.003637349s" podCreationTimestamp="2025-10-10 16:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:25:43.999679605 +0000 UTC m=+6046.449395153" watchObservedRunningTime="2025-10-10 16:25:44.003637349 +0000 UTC m=+6046.453352897" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.280983 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-gbd4d"] Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.286805 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.289227 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.293619 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.298625 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-gbd4d"] Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.467716 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-amphora-certs\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.467758 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/89b1209f-3797-42be-9f07-76b7ec120daf-config-data-merged\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.467877 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-scripts\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.467939 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-combined-ca-bundle\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.468003 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/89b1209f-3797-42be-9f07-76b7ec120daf-hm-ports\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.468042 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-config-data\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.569535 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-combined-ca-bundle\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.569613 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/89b1209f-3797-42be-9f07-76b7ec120daf-hm-ports\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.569654 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-config-data\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.569695 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-amphora-certs\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.569713 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/89b1209f-3797-42be-9f07-76b7ec120daf-config-data-merged\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.569773 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-scripts\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.571376 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/89b1209f-3797-42be-9f07-76b7ec120daf-config-data-merged\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.571732 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/89b1209f-3797-42be-9f07-76b7ec120daf-hm-ports\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.575650 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-scripts\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.577378 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-config-data\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.577934 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-amphora-certs\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.584777 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89b1209f-3797-42be-9f07-76b7ec120daf-combined-ca-bundle\") pod \"octavia-worker-gbd4d\" (UID: \"89b1209f-3797-42be-9f07-76b7ec120daf\") " pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.609805 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:44 crc kubenswrapper[4788]: I1010 16:25:44.991270 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-9n42v" event={"ID":"e74a446b-ff90-447e-82c4-a61d927dc0a3","Type":"ContainerStarted","Data":"332b5f0d6c942e223cce636fc42d7ba942bfdd1e9d0febd4de0a7719ad9745be"} Oct 10 16:25:45 crc kubenswrapper[4788]: I1010 16:25:45.184564 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-gbd4d"] Oct 10 16:25:46 crc kubenswrapper[4788]: I1010 16:25:46.000677 4788 generic.go:334] "Generic (PLEG): container finished" podID="e74a446b-ff90-447e-82c4-a61d927dc0a3" containerID="332b5f0d6c942e223cce636fc42d7ba942bfdd1e9d0febd4de0a7719ad9745be" exitCode=0 Oct 10 16:25:46 crc kubenswrapper[4788]: I1010 16:25:46.000721 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-9n42v" event={"ID":"e74a446b-ff90-447e-82c4-a61d927dc0a3","Type":"ContainerDied","Data":"332b5f0d6c942e223cce636fc42d7ba942bfdd1e9d0febd4de0a7719ad9745be"} Oct 10 16:25:46 crc kubenswrapper[4788]: I1010 16:25:46.002220 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-gbd4d" event={"ID":"89b1209f-3797-42be-9f07-76b7ec120daf","Type":"ContainerStarted","Data":"d9bf85792769101836072df0052e557830950e0722dad15235d3dc6276ebcdd5"} Oct 10 16:25:47 crc kubenswrapper[4788]: I1010 16:25:47.024797 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-9n42v" event={"ID":"e74a446b-ff90-447e-82c4-a61d927dc0a3","Type":"ContainerStarted","Data":"d661e34fd5b68e485c8ce1bc92cd2d206582273a337796a0b3fe966262377d2b"} Oct 10 16:25:47 crc kubenswrapper[4788]: I1010 16:25:47.025185 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:48 crc kubenswrapper[4788]: I1010 16:25:48.036561 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-gbd4d" event={"ID":"89b1209f-3797-42be-9f07-76b7ec120daf","Type":"ContainerStarted","Data":"05a58f1b1fe7a9cdb53c12a22971909f3a043bcabe4b870db01985674b9a7e21"} Oct 10 16:25:48 crc kubenswrapper[4788]: I1010 16:25:48.060885 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-9n42v" podStartSLOduration=5.946750241 podStartE2EDuration="8.060868445s" podCreationTimestamp="2025-10-10 16:25:40 +0000 UTC" firstStartedPulling="2025-10-10 16:25:41.689184925 +0000 UTC m=+6044.138900463" lastFinishedPulling="2025-10-10 16:25:43.803303109 +0000 UTC m=+6046.253018667" observedRunningTime="2025-10-10 16:25:47.050565642 +0000 UTC m=+6049.500281200" watchObservedRunningTime="2025-10-10 16:25:48.060868445 +0000 UTC m=+6050.510583993" Oct 10 16:25:49 crc kubenswrapper[4788]: I1010 16:25:49.047896 4788 generic.go:334] "Generic (PLEG): container finished" podID="89b1209f-3797-42be-9f07-76b7ec120daf" containerID="05a58f1b1fe7a9cdb53c12a22971909f3a043bcabe4b870db01985674b9a7e21" exitCode=0 Oct 10 16:25:49 crc kubenswrapper[4788]: I1010 16:25:49.047939 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-gbd4d" event={"ID":"89b1209f-3797-42be-9f07-76b7ec120daf","Type":"ContainerDied","Data":"05a58f1b1fe7a9cdb53c12a22971909f3a043bcabe4b870db01985674b9a7e21"} Oct 10 16:25:50 crc kubenswrapper[4788]: I1010 16:25:50.066274 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-gbd4d" event={"ID":"89b1209f-3797-42be-9f07-76b7ec120daf","Type":"ContainerStarted","Data":"d794ac2c20c6fb0f8ffae2f47cdf662c976c5b3e4996855673eca6a21e7d1b99"} Oct 10 16:25:50 crc kubenswrapper[4788]: I1010 16:25:50.066848 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-gbd4d" Oct 10 16:25:50 crc kubenswrapper[4788]: I1010 16:25:50.098437 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-gbd4d" podStartSLOduration=4.262231231 podStartE2EDuration="6.098416379s" podCreationTimestamp="2025-10-10 16:25:44 +0000 UTC" firstStartedPulling="2025-10-10 16:25:45.18657142 +0000 UTC m=+6047.636286968" lastFinishedPulling="2025-10-10 16:25:47.022756568 +0000 UTC m=+6049.472472116" observedRunningTime="2025-10-10 16:25:50.094558678 +0000 UTC m=+6052.544274226" watchObservedRunningTime="2025-10-10 16:25:50.098416379 +0000 UTC m=+6052.548131927" Oct 10 16:25:52 crc kubenswrapper[4788]: I1010 16:25:52.041368 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-m7hfp"] Oct 10 16:25:52 crc kubenswrapper[4788]: I1010 16:25:52.052248 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-m7hfp"] Oct 10 16:25:52 crc kubenswrapper[4788]: I1010 16:25:52.251073 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="585409fe-03f3-49b9-959b-6dffa35aa9ed" path="/var/lib/kubelet/pods/585409fe-03f3-49b9-959b-6dffa35aa9ed/volumes" Oct 10 16:25:54 crc kubenswrapper[4788]: I1010 16:25:54.217347 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-dlvvq" Oct 10 16:25:56 crc kubenswrapper[4788]: I1010 16:25:56.325686 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-9n42v" Oct 10 16:25:59 crc kubenswrapper[4788]: I1010 16:25:59.406478 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:25:59 crc kubenswrapper[4788]: I1010 16:25:59.406989 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:25:59 crc kubenswrapper[4788]: I1010 16:25:59.643665 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-gbd4d" Oct 10 16:26:02 crc kubenswrapper[4788]: I1010 16:26:02.028889 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6858-account-create-h5vvl"] Oct 10 16:26:02 crc kubenswrapper[4788]: I1010 16:26:02.039519 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6858-account-create-h5vvl"] Oct 10 16:26:02 crc kubenswrapper[4788]: I1010 16:26:02.244728 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7264913-5014-4307-a0eb-cd635b68a152" path="/var/lib/kubelet/pods/d7264913-5014-4307-a0eb-cd635b68a152/volumes" Oct 10 16:26:10 crc kubenswrapper[4788]: I1010 16:26:10.034407 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4x65t"] Oct 10 16:26:10 crc kubenswrapper[4788]: I1010 16:26:10.045767 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4x65t"] Oct 10 16:26:10 crc kubenswrapper[4788]: I1010 16:26:10.246915 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e118dac-a272-4a53-ac59-1ae9291402aa" path="/var/lib/kubelet/pods/1e118dac-a272-4a53-ac59-1ae9291402aa/volumes" Oct 10 16:26:13 crc kubenswrapper[4788]: I1010 16:26:13.831750 4788 scope.go:117] "RemoveContainer" containerID="055df83035bd2bd6fd838d33bfe726f548d044dafe230f490f21fb5bf0cd3d83" Oct 10 16:26:13 crc kubenswrapper[4788]: I1010 16:26:13.876978 4788 scope.go:117] "RemoveContainer" containerID="ff8ba4238c632907444b0341c93de733d6cd96167489db240b70f3285ac0b32f" Oct 10 16:26:13 crc kubenswrapper[4788]: I1010 16:26:13.903282 4788 scope.go:117] "RemoveContainer" containerID="dc551ed9855cf30e4f6fb3f25b37bfac335dd065e3a0713736c4806f4896ea20" Oct 10 16:26:13 crc kubenswrapper[4788]: I1010 16:26:13.948241 4788 scope.go:117] "RemoveContainer" containerID="23ed7bc67039c1344f91e2d4882f93ad0bb327feed0c8911b6809f25be4bca9f" Oct 10 16:26:13 crc kubenswrapper[4788]: I1010 16:26:13.969434 4788 scope.go:117] "RemoveContainer" containerID="960c4c7bf19c6b08d2b63577629db693d8befa34eead86a05bd574119e876057" Oct 10 16:26:14 crc kubenswrapper[4788]: I1010 16:26:14.025585 4788 scope.go:117] "RemoveContainer" containerID="cc3c3bd0ded678436d6110125c23a7051d8e416dde9d6e9417129781ad9f2763" Oct 10 16:26:14 crc kubenswrapper[4788]: I1010 16:26:14.044492 4788 scope.go:117] "RemoveContainer" containerID="77f1cb29ac193cced6eb4d49370d8b6d596e97131fdf59d267c0d4904374e209" Oct 10 16:26:29 crc kubenswrapper[4788]: I1010 16:26:29.406223 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:26:29 crc kubenswrapper[4788]: I1010 16:26:29.406749 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.043013 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k4vzx"] Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.048978 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.057903 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4vzx"] Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.102172 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7vnn\" (UniqueName: \"kubernetes.io/projected/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-kube-api-access-m7vnn\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.102263 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-catalog-content\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.102329 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-utilities\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.204232 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7vnn\" (UniqueName: \"kubernetes.io/projected/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-kube-api-access-m7vnn\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.204307 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-catalog-content\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.204351 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-utilities\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.204845 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-catalog-content\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.204898 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-utilities\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.227235 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7vnn\" (UniqueName: \"kubernetes.io/projected/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-kube-api-access-m7vnn\") pod \"community-operators-k4vzx\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.382312 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:41 crc kubenswrapper[4788]: I1010 16:26:41.903352 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4vzx"] Oct 10 16:26:42 crc kubenswrapper[4788]: I1010 16:26:42.638718 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerID="b8663e0286163c68306979205ec9c555fb8b98ad180df72891ebe93c6fbc4629" exitCode=0 Oct 10 16:26:42 crc kubenswrapper[4788]: I1010 16:26:42.638794 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4vzx" event={"ID":"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4","Type":"ContainerDied","Data":"b8663e0286163c68306979205ec9c555fb8b98ad180df72891ebe93c6fbc4629"} Oct 10 16:26:42 crc kubenswrapper[4788]: I1010 16:26:42.638834 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4vzx" event={"ID":"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4","Type":"ContainerStarted","Data":"221c3e157c54ff02cf6956937b6b66d8b73efb5caf42cc179b92bf011553372c"} Oct 10 16:26:42 crc kubenswrapper[4788]: I1010 16:26:42.642306 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:26:43 crc kubenswrapper[4788]: E1010 16:26:43.818785 4788 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.129:60664->38.102.83.129:41727: write tcp 38.102.83.129:60664->38.102.83.129:41727: write: broken pipe Oct 10 16:26:44 crc kubenswrapper[4788]: I1010 16:26:44.664610 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4vzx" event={"ID":"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4","Type":"ContainerStarted","Data":"6bd02580178594a07f070fa91403b84b174b7a73312773fc78d047e44df95558"} Oct 10 16:26:45 crc kubenswrapper[4788]: I1010 16:26:45.681131 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerID="6bd02580178594a07f070fa91403b84b174b7a73312773fc78d047e44df95558" exitCode=0 Oct 10 16:26:45 crc kubenswrapper[4788]: I1010 16:26:45.681244 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4vzx" event={"ID":"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4","Type":"ContainerDied","Data":"6bd02580178594a07f070fa91403b84b174b7a73312773fc78d047e44df95558"} Oct 10 16:26:46 crc kubenswrapper[4788]: I1010 16:26:46.702371 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4vzx" event={"ID":"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4","Type":"ContainerStarted","Data":"c6fa78e9732614efa7b5b74896672e01e846b87cdceb0cd3d2bb2e64ebbcaf2d"} Oct 10 16:26:46 crc kubenswrapper[4788]: I1010 16:26:46.745223 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k4vzx" podStartSLOduration=2.268523013 podStartE2EDuration="5.745185573s" podCreationTimestamp="2025-10-10 16:26:41 +0000 UTC" firstStartedPulling="2025-10-10 16:26:42.641619023 +0000 UTC m=+6105.091334621" lastFinishedPulling="2025-10-10 16:26:46.118281623 +0000 UTC m=+6108.567997181" observedRunningTime="2025-10-10 16:26:46.730296731 +0000 UTC m=+6109.180012299" watchObservedRunningTime="2025-10-10 16:26:46.745185573 +0000 UTC m=+6109.194901151" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.069311 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67f4c99c4f-6cnmx"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.073479 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.077342 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.077342 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-bppnp" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.077349 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.078439 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.088511 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67f4c99c4f-6cnmx"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.089842 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c792f5-3274-4ff6-b87b-942398c3671d-logs\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.089912 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/29c792f5-3274-4ff6-b87b-942398c3671d-horizon-secret-key\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.089994 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-config-data\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.090047 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql725\" (UniqueName: \"kubernetes.io/projected/29c792f5-3274-4ff6-b87b-942398c3671d-kube-api-access-ql725\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.090074 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-scripts\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.143294 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.143630 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerName="glance-log" containerID="cri-o://ea00096efb6dc90a7bad6d9e9a7da803eccf8477e55a62dc236a87f44aa69e1e" gracePeriod=30 Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.144166 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerName="glance-httpd" containerID="cri-o://8c13c7adb5a1a8363332b520e23b991f39b08314bc744aa9f43ca67d699afda5" gracePeriod=30 Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.191357 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c792f5-3274-4ff6-b87b-942398c3671d-logs\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.191410 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/29c792f5-3274-4ff6-b87b-942398c3671d-horizon-secret-key\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.191479 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-config-data\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.191523 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql725\" (UniqueName: \"kubernetes.io/projected/29c792f5-3274-4ff6-b87b-942398c3671d-kube-api-access-ql725\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.191543 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-scripts\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.192725 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-scripts\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.193754 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c792f5-3274-4ff6-b87b-942398c3671d-logs\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.200903 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-config-data\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.206601 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/29c792f5-3274-4ff6-b87b-942398c3671d-horizon-secret-key\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.220212 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql725\" (UniqueName: \"kubernetes.io/projected/29c792f5-3274-4ff6-b87b-942398c3671d-kube-api-access-ql725\") pod \"horizon-67f4c99c4f-6cnmx\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.228748 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67cb58f66f-8p4t2"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.231058 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.248926 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.249569 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-log" containerID="cri-o://39c96d7e490e032971909cbaf3a95df7e21aaf7fa95cc28e68d4cb76fc8a5c7a" gracePeriod=30 Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.249848 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-httpd" containerID="cri-o://7eae1f41b35d6cd12e721bea2a731f1610c78b398434a5a866d429fe7716e144" gracePeriod=30 Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.285696 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67cb58f66f-8p4t2"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.292979 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-config-data\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.293066 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlvpz\" (UniqueName: \"kubernetes.io/projected/86ca1055-1050-4d17-8de6-e74a3d72eafc-kube-api-access-wlvpz\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.293179 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86ca1055-1050-4d17-8de6-e74a3d72eafc-logs\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.293219 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-scripts\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.293281 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/86ca1055-1050-4d17-8de6-e74a3d72eafc-horizon-secret-key\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.394890 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-scripts\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.394946 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/86ca1055-1050-4d17-8de6-e74a3d72eafc-horizon-secret-key\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.395055 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-config-data\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.395095 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlvpz\" (UniqueName: \"kubernetes.io/projected/86ca1055-1050-4d17-8de6-e74a3d72eafc-kube-api-access-wlvpz\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.395182 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86ca1055-1050-4d17-8de6-e74a3d72eafc-logs\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.395903 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-scripts\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.396402 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86ca1055-1050-4d17-8de6-e74a3d72eafc-logs\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.400066 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-config-data\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.402750 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/86ca1055-1050-4d17-8de6-e74a3d72eafc-horizon-secret-key\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.416007 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlvpz\" (UniqueName: \"kubernetes.io/projected/86ca1055-1050-4d17-8de6-e74a3d72eafc-kube-api-access-wlvpz\") pod \"horizon-67cb58f66f-8p4t2\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.444915 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.451491 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.732072 4788 generic.go:334] "Generic (PLEG): container finished" podID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerID="ea00096efb6dc90a7bad6d9e9a7da803eccf8477e55a62dc236a87f44aa69e1e" exitCode=143 Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.732187 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12d32477-f9af-47c5-9b91-0ab82d677e5a","Type":"ContainerDied","Data":"ea00096efb6dc90a7bad6d9e9a7da803eccf8477e55a62dc236a87f44aa69e1e"} Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.740135 4788 generic.go:334] "Generic (PLEG): container finished" podID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerID="39c96d7e490e032971909cbaf3a95df7e21aaf7fa95cc28e68d4cb76fc8a5c7a" exitCode=143 Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.740198 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ce16116b-c7d1-4d82-ab72-34002a0dbecc","Type":"ContainerDied","Data":"39c96d7e490e032971909cbaf3a95df7e21aaf7fa95cc28e68d4cb76fc8a5c7a"} Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.826348 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67cb58f66f-8p4t2"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.871709 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-65cbd96cf7-rw4r7"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.874008 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.885578 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65cbd96cf7-rw4r7"] Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.906177 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j52z\" (UniqueName: \"kubernetes.io/projected/a32e043f-56f6-4468-b2d5-ef00f3854626-kube-api-access-9j52z\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.906278 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32e043f-56f6-4468-b2d5-ef00f3854626-logs\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.906334 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a32e043f-56f6-4468-b2d5-ef00f3854626-horizon-secret-key\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.906410 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-scripts\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.906426 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-config-data\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.983671 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67f4c99c4f-6cnmx"] Oct 10 16:26:48 crc kubenswrapper[4788]: W1010 16:26:48.985296 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29c792f5_3274_4ff6_b87b_942398c3671d.slice/crio-7c6e7daf22b03423e67261f8b5f91b6ec094f3154a567e17735d5bd2f7ecb2c5 WatchSource:0}: Error finding container 7c6e7daf22b03423e67261f8b5f91b6ec094f3154a567e17735d5bd2f7ecb2c5: Status 404 returned error can't find the container with id 7c6e7daf22b03423e67261f8b5f91b6ec094f3154a567e17735d5bd2f7ecb2c5 Oct 10 16:26:48 crc kubenswrapper[4788]: I1010 16:26:48.995312 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67cb58f66f-8p4t2"] Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.007861 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j52z\" (UniqueName: \"kubernetes.io/projected/a32e043f-56f6-4468-b2d5-ef00f3854626-kube-api-access-9j52z\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.007956 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32e043f-56f6-4468-b2d5-ef00f3854626-logs\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.008017 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a32e043f-56f6-4468-b2d5-ef00f3854626-horizon-secret-key\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.008078 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-scripts\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.008103 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-config-data\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.009746 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32e043f-56f6-4468-b2d5-ef00f3854626-logs\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.010000 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-scripts\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.011022 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-config-data\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.017905 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a32e043f-56f6-4468-b2d5-ef00f3854626-horizon-secret-key\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.029702 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j52z\" (UniqueName: \"kubernetes.io/projected/a32e043f-56f6-4468-b2d5-ef00f3854626-kube-api-access-9j52z\") pod \"horizon-65cbd96cf7-rw4r7\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.207970 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.720794 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65cbd96cf7-rw4r7"] Oct 10 16:26:49 crc kubenswrapper[4788]: W1010 16:26:49.730573 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda32e043f_56f6_4468_b2d5_ef00f3854626.slice/crio-77011c0216e8ae9f5f141145a4fd20363e773ea4ccefaeea6380a26020dd5bfc WatchSource:0}: Error finding container 77011c0216e8ae9f5f141145a4fd20363e773ea4ccefaeea6380a26020dd5bfc: Status 404 returned error can't find the container with id 77011c0216e8ae9f5f141145a4fd20363e773ea4ccefaeea6380a26020dd5bfc Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.753033 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f4c99c4f-6cnmx" event={"ID":"29c792f5-3274-4ff6-b87b-942398c3671d","Type":"ContainerStarted","Data":"7c6e7daf22b03423e67261f8b5f91b6ec094f3154a567e17735d5bd2f7ecb2c5"} Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.756118 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65cbd96cf7-rw4r7" event={"ID":"a32e043f-56f6-4468-b2d5-ef00f3854626","Type":"ContainerStarted","Data":"77011c0216e8ae9f5f141145a4fd20363e773ea4ccefaeea6380a26020dd5bfc"} Oct 10 16:26:49 crc kubenswrapper[4788]: I1010 16:26:49.758196 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb58f66f-8p4t2" event={"ID":"86ca1055-1050-4d17-8de6-e74a3d72eafc","Type":"ContainerStarted","Data":"2266d455cf89c2810060c440b14406aba5aa81f8dd18e1412c9e348eadc0d38b"} Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.383394 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.388413 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.441047 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.677345 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.1.49:9292/healthcheck\": read tcp 10.217.0.2:39516->10.217.1.49:9292: read: connection reset by peer" Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.677440 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.1.49:9292/healthcheck\": read tcp 10.217.0.2:39504->10.217.1.49:9292: read: connection reset by peer" Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.786558 4788 generic.go:334] "Generic (PLEG): container finished" podID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerID="8c13c7adb5a1a8363332b520e23b991f39b08314bc744aa9f43ca67d699afda5" exitCode=0 Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.786611 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12d32477-f9af-47c5-9b91-0ab82d677e5a","Type":"ContainerDied","Data":"8c13c7adb5a1a8363332b520e23b991f39b08314bc744aa9f43ca67d699afda5"} Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.790982 4788 generic.go:334] "Generic (PLEG): container finished" podID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerID="7eae1f41b35d6cd12e721bea2a731f1610c78b398434a5a866d429fe7716e144" exitCode=0 Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.791077 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ce16116b-c7d1-4d82-ab72-34002a0dbecc","Type":"ContainerDied","Data":"7eae1f41b35d6cd12e721bea2a731f1610c78b398434a5a866d429fe7716e144"} Oct 10 16:26:51 crc kubenswrapper[4788]: I1010 16:26:51.837933 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:53 crc kubenswrapper[4788]: I1010 16:26:53.041822 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6ndzr"] Oct 10 16:26:53 crc kubenswrapper[4788]: I1010 16:26:53.055094 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6ndzr"] Oct 10 16:26:54 crc kubenswrapper[4788]: I1010 16:26:54.248782 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e755581d-b00e-41d1-bb31-e4384bbf444f" path="/var/lib/kubelet/pods/e755581d-b00e-41d1-bb31-e4384bbf444f/volumes" Oct 10 16:26:54 crc kubenswrapper[4788]: I1010 16:26:54.252621 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k4vzx"] Oct 10 16:26:54 crc kubenswrapper[4788]: I1010 16:26:54.253105 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k4vzx" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerName="registry-server" containerID="cri-o://c6fa78e9732614efa7b5b74896672e01e846b87cdceb0cd3d2bb2e64ebbcaf2d" gracePeriod=2 Oct 10 16:26:54 crc kubenswrapper[4788]: I1010 16:26:54.829690 4788 generic.go:334] "Generic (PLEG): container finished" podID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerID="c6fa78e9732614efa7b5b74896672e01e846b87cdceb0cd3d2bb2e64ebbcaf2d" exitCode=0 Oct 10 16:26:54 crc kubenswrapper[4788]: I1010 16:26:54.829744 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4vzx" event={"ID":"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4","Type":"ContainerDied","Data":"c6fa78e9732614efa7b5b74896672e01e846b87cdceb0cd3d2bb2e64ebbcaf2d"} Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.188051 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.233016 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-catalog-content\") pod \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.237901 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7vnn\" (UniqueName: \"kubernetes.io/projected/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-kube-api-access-m7vnn\") pod \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.238369 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-utilities\") pod \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\" (UID: \"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.239296 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-utilities" (OuterVolumeSpecName: "utilities") pod "e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" (UID: "e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.245627 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-kube-api-access-m7vnn" (OuterVolumeSpecName: "kube-api-access-m7vnn") pod "e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" (UID: "e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4"). InnerVolumeSpecName "kube-api-access-m7vnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.265889 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.291702 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" (UID: "e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.339972 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-httpd-run\") pod \"12d32477-f9af-47c5-9b91-0ab82d677e5a\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.340028 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-scripts\") pod \"12d32477-f9af-47c5-9b91-0ab82d677e5a\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.340293 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-config-data\") pod \"12d32477-f9af-47c5-9b91-0ab82d677e5a\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.340758 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq7s9\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-kube-api-access-kq7s9\") pod \"12d32477-f9af-47c5-9b91-0ab82d677e5a\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.341330 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-logs\") pod \"12d32477-f9af-47c5-9b91-0ab82d677e5a\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.341415 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-ceph\") pod \"12d32477-f9af-47c5-9b91-0ab82d677e5a\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.341514 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-combined-ca-bundle\") pod \"12d32477-f9af-47c5-9b91-0ab82d677e5a\" (UID: \"12d32477-f9af-47c5-9b91-0ab82d677e5a\") " Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.342064 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.342076 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.342087 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7vnn\" (UniqueName: \"kubernetes.io/projected/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4-kube-api-access-m7vnn\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.350626 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-logs" (OuterVolumeSpecName: "logs") pod "12d32477-f9af-47c5-9b91-0ab82d677e5a" (UID: "12d32477-f9af-47c5-9b91-0ab82d677e5a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.350763 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "12d32477-f9af-47c5-9b91-0ab82d677e5a" (UID: "12d32477-f9af-47c5-9b91-0ab82d677e5a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.362034 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-scripts" (OuterVolumeSpecName: "scripts") pod "12d32477-f9af-47c5-9b91-0ab82d677e5a" (UID: "12d32477-f9af-47c5-9b91-0ab82d677e5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.364450 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-kube-api-access-kq7s9" (OuterVolumeSpecName: "kube-api-access-kq7s9") pod "12d32477-f9af-47c5-9b91-0ab82d677e5a" (UID: "12d32477-f9af-47c5-9b91-0ab82d677e5a"). InnerVolumeSpecName "kube-api-access-kq7s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.364793 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-ceph" (OuterVolumeSpecName: "ceph") pod "12d32477-f9af-47c5-9b91-0ab82d677e5a" (UID: "12d32477-f9af-47c5-9b91-0ab82d677e5a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.411401 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12d32477-f9af-47c5-9b91-0ab82d677e5a" (UID: "12d32477-f9af-47c5-9b91-0ab82d677e5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.440299 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-config-data" (OuterVolumeSpecName: "config-data") pod "12d32477-f9af-47c5-9b91-0ab82d677e5a" (UID: "12d32477-f9af-47c5-9b91-0ab82d677e5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.444626 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.444849 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.444910 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.444996 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d32477-f9af-47c5-9b91-0ab82d677e5a-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.445056 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq7s9\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-kube-api-access-kq7s9\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.445124 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d32477-f9af-47c5-9b91-0ab82d677e5a-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:56 crc kubenswrapper[4788]: I1010 16:26:56.445219 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/12d32477-f9af-47c5-9b91-0ab82d677e5a-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.862104 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb58f66f-8p4t2" event={"ID":"86ca1055-1050-4d17-8de6-e74a3d72eafc","Type":"ContainerStarted","Data":"815417e3409a3ffb3d8bd21f037b20c1cbc6aa030eb6bad68641d161f1a05ae4"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.862425 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb58f66f-8p4t2" event={"ID":"86ca1055-1050-4d17-8de6-e74a3d72eafc","Type":"ContainerStarted","Data":"ae9c05815d38508e24dbac6c4f6c1270ac67bf7cf118b5523ebe04760f389751"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.862275 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67cb58f66f-8p4t2" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerName="horizon" containerID="cri-o://815417e3409a3ffb3d8bd21f037b20c1cbc6aa030eb6bad68641d161f1a05ae4" gracePeriod=30 Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.862239 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67cb58f66f-8p4t2" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerName="horizon-log" containerID="cri-o://ae9c05815d38508e24dbac6c4f6c1270ac67bf7cf118b5523ebe04760f389751" gracePeriod=30 Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.867150 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4vzx" event={"ID":"e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4","Type":"ContainerDied","Data":"221c3e157c54ff02cf6956937b6b66d8b73efb5caf42cc179b92bf011553372c"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.867207 4788 scope.go:117] "RemoveContainer" containerID="c6fa78e9732614efa7b5b74896672e01e846b87cdceb0cd3d2bb2e64ebbcaf2d" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.867335 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4vzx" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.878934 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f4c99c4f-6cnmx" event={"ID":"29c792f5-3274-4ff6-b87b-942398c3671d","Type":"ContainerStarted","Data":"22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.878980 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f4c99c4f-6cnmx" event={"ID":"29c792f5-3274-4ff6-b87b-942398c3671d","Type":"ContainerStarted","Data":"13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.886664 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65cbd96cf7-rw4r7" event={"ID":"a32e043f-56f6-4468-b2d5-ef00f3854626","Type":"ContainerStarted","Data":"e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.886744 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65cbd96cf7-rw4r7" event={"ID":"a32e043f-56f6-4468-b2d5-ef00f3854626","Type":"ContainerStarted","Data":"6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.905750 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67cb58f66f-8p4t2" podStartSLOduration=2.063260686 podStartE2EDuration="8.905721576s" podCreationTimestamp="2025-10-10 16:26:48 +0000 UTC" firstStartedPulling="2025-10-10 16:26:48.993453491 +0000 UTC m=+6111.443169039" lastFinishedPulling="2025-10-10 16:26:55.835914381 +0000 UTC m=+6118.285629929" observedRunningTime="2025-10-10 16:26:56.902883011 +0000 UTC m=+6119.352598559" watchObservedRunningTime="2025-10-10 16:26:56.905721576 +0000 UTC m=+6119.355437114" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.907485 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12d32477-f9af-47c5-9b91-0ab82d677e5a","Type":"ContainerDied","Data":"93db2822c04ee7d25781f3dbe070716cdd7ab2920a28fd60f505b14c397f3d21"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.907580 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.934087 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-65cbd96cf7-rw4r7" podStartSLOduration=2.811482019 podStartE2EDuration="8.934071084s" podCreationTimestamp="2025-10-10 16:26:48 +0000 UTC" firstStartedPulling="2025-10-10 16:26:49.73316282 +0000 UTC m=+6112.182878368" lastFinishedPulling="2025-10-10 16:26:55.855751845 +0000 UTC m=+6118.305467433" observedRunningTime="2025-10-10 16:26:56.928057875 +0000 UTC m=+6119.377773423" watchObservedRunningTime="2025-10-10 16:26:56.934071084 +0000 UTC m=+6119.383786632" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.934318 4788 scope.go:117] "RemoveContainer" containerID="6bd02580178594a07f070fa91403b84b174b7a73312773fc78d047e44df95558" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:56.959097 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67f4c99c4f-6cnmx" podStartSLOduration=2.131515127 podStartE2EDuration="8.959075064s" podCreationTimestamp="2025-10-10 16:26:48 +0000 UTC" firstStartedPulling="2025-10-10 16:26:48.988297455 +0000 UTC m=+6111.438013003" lastFinishedPulling="2025-10-10 16:26:55.815857372 +0000 UTC m=+6118.265572940" observedRunningTime="2025-10-10 16:26:56.954539975 +0000 UTC m=+6119.404255523" watchObservedRunningTime="2025-10-10 16:26:56.959075064 +0000 UTC m=+6119.408790612" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.004331 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k4vzx"] Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.011889 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.016721 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k4vzx"] Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.020788 4788 scope.go:117] "RemoveContainer" containerID="b8663e0286163c68306979205ec9c555fb8b98ad180df72891ebe93c6fbc4629" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.057276 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-ceph\") pod \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.057466 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-config-data\") pod \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.057551 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-logs\") pod \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.057599 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-scripts\") pod \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.057622 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g2hf\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-kube-api-access-9g2hf\") pod \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.057647 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-combined-ca-bundle\") pod \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.057665 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-httpd-run\") pod \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\" (UID: \"ce16116b-c7d1-4d82-ab72-34002a0dbecc\") " Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.063060 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-logs" (OuterVolumeSpecName: "logs") pod "ce16116b-c7d1-4d82-ab72-34002a0dbecc" (UID: "ce16116b-c7d1-4d82-ab72-34002a0dbecc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.063513 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ce16116b-c7d1-4d82-ab72-34002a0dbecc" (UID: "ce16116b-c7d1-4d82-ab72-34002a0dbecc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.080538 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-ceph" (OuterVolumeSpecName: "ceph") pod "ce16116b-c7d1-4d82-ab72-34002a0dbecc" (UID: "ce16116b-c7d1-4d82-ab72-34002a0dbecc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.082836 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-kube-api-access-9g2hf" (OuterVolumeSpecName: "kube-api-access-9g2hf") pod "ce16116b-c7d1-4d82-ab72-34002a0dbecc" (UID: "ce16116b-c7d1-4d82-ab72-34002a0dbecc"). InnerVolumeSpecName "kube-api-access-9g2hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.112867 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-scripts" (OuterVolumeSpecName: "scripts") pod "ce16116b-c7d1-4d82-ab72-34002a0dbecc" (UID: "ce16116b-c7d1-4d82-ab72-34002a0dbecc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.126653 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.129394 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce16116b-c7d1-4d82-ab72-34002a0dbecc" (UID: "ce16116b-c7d1-4d82-ab72-34002a0dbecc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.143900 4788 scope.go:117] "RemoveContainer" containerID="8c13c7adb5a1a8363332b520e23b991f39b08314bc744aa9f43ca67d699afda5" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.166323 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.166998 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.167039 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.167055 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g2hf\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-kube-api-access-9g2hf\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.167068 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.167080 4788 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce16116b-c7d1-4d82-ab72-34002a0dbecc-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.167095 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ce16116b-c7d1-4d82-ab72-34002a0dbecc-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.170264 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-config-data" (OuterVolumeSpecName: "config-data") pod "ce16116b-c7d1-4d82-ab72-34002a0dbecc" (UID: "ce16116b-c7d1-4d82-ab72-34002a0dbecc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.177017 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:26:57 crc kubenswrapper[4788]: E1010 16:26:57.177536 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerName="glance-httpd" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.177550 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerName="glance-httpd" Oct 10 16:26:57 crc kubenswrapper[4788]: E1010 16:26:57.177568 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerName="extract-content" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.177574 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerName="extract-content" Oct 10 16:26:57 crc kubenswrapper[4788]: E1010 16:26:57.177588 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-log" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.177596 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-log" Oct 10 16:26:57 crc kubenswrapper[4788]: E1010 16:26:57.177622 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-httpd" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.177628 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-httpd" Oct 10 16:26:57 crc kubenswrapper[4788]: E1010 16:26:57.177640 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerName="registry-server" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.177646 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerName="registry-server" Oct 10 16:26:57 crc kubenswrapper[4788]: E1010 16:26:57.177659 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerName="glance-log" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.177666 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerName="glance-log" Oct 10 16:26:57 crc kubenswrapper[4788]: E1010 16:26:57.177691 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerName="extract-utilities" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.177697 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerName="extract-utilities" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.196071 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-log" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.197641 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerName="glance-httpd" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.197659 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" containerName="registry-server" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.197716 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" containerName="glance-log" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.197754 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" containerName="glance-httpd" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.208671 4788 scope.go:117] "RemoveContainer" containerID="ea00096efb6dc90a7bad6d9e9a7da803eccf8477e55a62dc236a87f44aa69e1e" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.220561 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.220789 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.224387 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.268679 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d65afb02-bfb8-4467-a722-1630581e302c-logs\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.268758 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d65afb02-bfb8-4467-a722-1630581e302c-ceph\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.268833 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-scripts\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.268860 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d65afb02-bfb8-4467-a722-1630581e302c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.268880 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.268956 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wptf6\" (UniqueName: \"kubernetes.io/projected/d65afb02-bfb8-4467-a722-1630581e302c-kube-api-access-wptf6\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.268999 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-config-data\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.269083 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce16116b-c7d1-4d82-ab72-34002a0dbecc-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.370596 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-config-data\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.370688 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d65afb02-bfb8-4467-a722-1630581e302c-logs\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.370749 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d65afb02-bfb8-4467-a722-1630581e302c-ceph\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.370796 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-scripts\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.370819 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d65afb02-bfb8-4467-a722-1630581e302c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.370836 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.370900 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wptf6\" (UniqueName: \"kubernetes.io/projected/d65afb02-bfb8-4467-a722-1630581e302c-kube-api-access-wptf6\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.371626 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d65afb02-bfb8-4467-a722-1630581e302c-logs\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.372662 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d65afb02-bfb8-4467-a722-1630581e302c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.382583 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-scripts\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.382820 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d65afb02-bfb8-4467-a722-1630581e302c-ceph\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.382847 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.382856 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d65afb02-bfb8-4467-a722-1630581e302c-config-data\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.388076 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wptf6\" (UniqueName: \"kubernetes.io/projected/d65afb02-bfb8-4467-a722-1630581e302c-kube-api-access-wptf6\") pod \"glance-default-external-api-0\" (UID: \"d65afb02-bfb8-4467-a722-1630581e302c\") " pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.584671 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.920106 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ce16116b-c7d1-4d82-ab72-34002a0dbecc","Type":"ContainerDied","Data":"399b0465c36ca0ae9ee6c85416552c8749fc16ef3d88b765317e7f62a8acabe6"} Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.920419 4788 scope.go:117] "RemoveContainer" containerID="7eae1f41b35d6cd12e721bea2a731f1610c78b398434a5a866d429fe7716e144" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.920167 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.961513 4788 scope.go:117] "RemoveContainer" containerID="39c96d7e490e032971909cbaf3a95df7e21aaf7fa95cc28e68d4cb76fc8a5c7a" Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.969120 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.985250 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.995799 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:26:57 crc kubenswrapper[4788]: I1010 16:26:57.998236 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.004611 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.021436 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.167740 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.200820 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31093023-af84-44fa-a6e7-a112b823b67c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.200906 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9th6\" (UniqueName: \"kubernetes.io/projected/31093023-af84-44fa-a6e7-a112b823b67c-kube-api-access-s9th6\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.200942 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/31093023-af84-44fa-a6e7-a112b823b67c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.200962 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.201008 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.201031 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.201065 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31093023-af84-44fa-a6e7-a112b823b67c-logs\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.254709 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12d32477-f9af-47c5-9b91-0ab82d677e5a" path="/var/lib/kubelet/pods/12d32477-f9af-47c5-9b91-0ab82d677e5a/volumes" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.255623 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce16116b-c7d1-4d82-ab72-34002a0dbecc" path="/var/lib/kubelet/pods/ce16116b-c7d1-4d82-ab72-34002a0dbecc/volumes" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.256749 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4" path="/var/lib/kubelet/pods/e9c3c7d7-8882-40d3-a0bf-fb09d7a641c4/volumes" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.302502 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.302732 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.302847 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31093023-af84-44fa-a6e7-a112b823b67c-logs\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.302980 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31093023-af84-44fa-a6e7-a112b823b67c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.303104 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9th6\" (UniqueName: \"kubernetes.io/projected/31093023-af84-44fa-a6e7-a112b823b67c-kube-api-access-s9th6\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.303286 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/31093023-af84-44fa-a6e7-a112b823b67c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.303370 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.303516 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31093023-af84-44fa-a6e7-a112b823b67c-logs\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.303856 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31093023-af84-44fa-a6e7-a112b823b67c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.317334 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.326786 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/31093023-af84-44fa-a6e7-a112b823b67c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.327484 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.329094 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.331690 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31093023-af84-44fa-a6e7-a112b823b67c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.344833 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9th6\" (UniqueName: \"kubernetes.io/projected/31093023-af84-44fa-a6e7-a112b823b67c-kube-api-access-s9th6\") pod \"glance-default-internal-api-0\" (UID: \"31093023-af84-44fa-a6e7-a112b823b67c\") " pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.445294 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.445382 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.453213 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.636079 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 10 16:26:58 crc kubenswrapper[4788]: I1010 16:26:58.940908 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d65afb02-bfb8-4467-a722-1630581e302c","Type":"ContainerStarted","Data":"745f77a978893b0cf3cf075c9685c9c139562b6130e33d391b60046261f25402"} Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.196600 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.209100 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.209324 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:26:59 crc kubenswrapper[4788]: W1010 16:26:59.211354 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31093023_af84_44fa_a6e7_a112b823b67c.slice/crio-5c7d64d5b22b247926475f5e95baa2a234121d277b135cc6c48d59bc0fc07272 WatchSource:0}: Error finding container 5c7d64d5b22b247926475f5e95baa2a234121d277b135cc6c48d59bc0fc07272: Status 404 returned error can't find the container with id 5c7d64d5b22b247926475f5e95baa2a234121d277b135cc6c48d59bc0fc07272 Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.406182 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.406251 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.406309 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.407094 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e510a005acb06733b8f662e2f362c13e2ba92b9f626ca5375903dde617f1f2fe"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.407177 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://e510a005acb06733b8f662e2f362c13e2ba92b9f626ca5375903dde617f1f2fe" gracePeriod=600 Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.973406 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="e510a005acb06733b8f662e2f362c13e2ba92b9f626ca5375903dde617f1f2fe" exitCode=0 Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.973934 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"e510a005acb06733b8f662e2f362c13e2ba92b9f626ca5375903dde617f1f2fe"} Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.974272 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3"} Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.974301 4788 scope.go:117] "RemoveContainer" containerID="5456995bba5cb460d01a76224a264f3c9485db111b12ebc8e414596fe4a6a1c2" Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.982418 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d65afb02-bfb8-4467-a722-1630581e302c","Type":"ContainerStarted","Data":"5676589af28ed92a6951dfb54e6f454f92f2ffcbeff1efc99c7c9d579a2b828a"} Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.982469 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d65afb02-bfb8-4467-a722-1630581e302c","Type":"ContainerStarted","Data":"acb354d860be6f398e7b8594f147ef034d48af04764a121a8312fecf12bb7122"} Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.985109 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31093023-af84-44fa-a6e7-a112b823b67c","Type":"ContainerStarted","Data":"de7a27a0c92e29afaf91082d96ea2c8acb33d7746938827b4c428e59807efd05"} Oct 10 16:26:59 crc kubenswrapper[4788]: I1010 16:26:59.985236 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31093023-af84-44fa-a6e7-a112b823b67c","Type":"ContainerStarted","Data":"5c7d64d5b22b247926475f5e95baa2a234121d277b135cc6c48d59bc0fc07272"} Oct 10 16:27:00 crc kubenswrapper[4788]: I1010 16:27:00.019400 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.019381841 podStartE2EDuration="3.019381841s" podCreationTimestamp="2025-10-10 16:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:27:00.01860752 +0000 UTC m=+6122.468323068" watchObservedRunningTime="2025-10-10 16:27:00.019381841 +0000 UTC m=+6122.469097389" Oct 10 16:27:01 crc kubenswrapper[4788]: I1010 16:27:01.015879 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31093023-af84-44fa-a6e7-a112b823b67c","Type":"ContainerStarted","Data":"ede69cbc1db2c6891d9a9c2e72e97305df447ea5bef9e3718d9d50ba2a1a6692"} Oct 10 16:27:01 crc kubenswrapper[4788]: I1010 16:27:01.059992 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.059966334 podStartE2EDuration="4.059966334s" podCreationTimestamp="2025-10-10 16:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:27:01.046808007 +0000 UTC m=+6123.496523575" watchObservedRunningTime="2025-10-10 16:27:01.059966334 +0000 UTC m=+6123.509681892" Oct 10 16:27:03 crc kubenswrapper[4788]: I1010 16:27:03.040006 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-9279-account-create-6hfhf"] Oct 10 16:27:03 crc kubenswrapper[4788]: I1010 16:27:03.053699 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-9279-account-create-6hfhf"] Oct 10 16:27:04 crc kubenswrapper[4788]: I1010 16:27:04.248594 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="578cd6ee-08cf-4a90-bb89-f9babce23451" path="/var/lib/kubelet/pods/578cd6ee-08cf-4a90-bb89-f9babce23451/volumes" Oct 10 16:27:07 crc kubenswrapper[4788]: I1010 16:27:07.584886 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 16:27:07 crc kubenswrapper[4788]: I1010 16:27:07.585398 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 10 16:27:07 crc kubenswrapper[4788]: I1010 16:27:07.614957 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 16:27:07 crc kubenswrapper[4788]: I1010 16:27:07.638683 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 10 16:27:08 crc kubenswrapper[4788]: I1010 16:27:08.086779 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 16:27:08 crc kubenswrapper[4788]: I1010 16:27:08.086827 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 10 16:27:08 crc kubenswrapper[4788]: I1010 16:27:08.448201 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-67f4c99c4f-6cnmx" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 10 16:27:08 crc kubenswrapper[4788]: I1010 16:27:08.637218 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 16:27:08 crc kubenswrapper[4788]: I1010 16:27:08.637278 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 10 16:27:08 crc kubenswrapper[4788]: I1010 16:27:08.683357 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 16:27:08 crc kubenswrapper[4788]: I1010 16:27:08.688046 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 10 16:27:09 crc kubenswrapper[4788]: I1010 16:27:09.102791 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 16:27:09 crc kubenswrapper[4788]: I1010 16:27:09.103419 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 10 16:27:09 crc kubenswrapper[4788]: I1010 16:27:09.212907 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-65cbd96cf7-rw4r7" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Oct 10 16:27:10 crc kubenswrapper[4788]: I1010 16:27:10.196599 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 16:27:10 crc kubenswrapper[4788]: I1010 16:27:10.197030 4788 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 16:27:10 crc kubenswrapper[4788]: I1010 16:27:10.334027 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 10 16:27:11 crc kubenswrapper[4788]: I1010 16:27:11.253844 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 16:27:11 crc kubenswrapper[4788]: I1010 16:27:11.254389 4788 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 10 16:27:11 crc kubenswrapper[4788]: I1010 16:27:11.286216 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 10 16:27:12 crc kubenswrapper[4788]: I1010 16:27:12.034517 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-j58jf"] Oct 10 16:27:12 crc kubenswrapper[4788]: I1010 16:27:12.045862 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-j58jf"] Oct 10 16:27:12 crc kubenswrapper[4788]: I1010 16:27:12.249880 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f72954d6-f07f-4272-a985-4dbcc0ae7855" path="/var/lib/kubelet/pods/f72954d6-f07f-4272-a985-4dbcc0ae7855/volumes" Oct 10 16:27:14 crc kubenswrapper[4788]: I1010 16:27:14.184902 4788 scope.go:117] "RemoveContainer" containerID="8d74f14f749a6965f4b498bdd02e32e5ff93ef766be7eccc545285f7c8808842" Oct 10 16:27:14 crc kubenswrapper[4788]: I1010 16:27:14.217251 4788 scope.go:117] "RemoveContainer" containerID="f4331ca8c2fa2441de223d36d991407d7f1588ed8124c6e47353b5868391dfc7" Oct 10 16:27:14 crc kubenswrapper[4788]: I1010 16:27:14.293450 4788 scope.go:117] "RemoveContainer" containerID="38a1d75692076941e8b31beca56563bb4fd1cb9e8b493f305f28a2496eb5cbc2" Oct 10 16:27:20 crc kubenswrapper[4788]: I1010 16:27:20.218897 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:27:21 crc kubenswrapper[4788]: I1010 16:27:21.023335 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:27:21 crc kubenswrapper[4788]: I1010 16:27:21.876959 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:27:22 crc kubenswrapper[4788]: I1010 16:27:22.732745 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:27:22 crc kubenswrapper[4788]: I1010 16:27:22.793173 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67f4c99c4f-6cnmx"] Oct 10 16:27:22 crc kubenswrapper[4788]: I1010 16:27:22.793500 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67f4c99c4f-6cnmx" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon-log" containerID="cri-o://13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2" gracePeriod=30 Oct 10 16:27:22 crc kubenswrapper[4788]: I1010 16:27:22.793913 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67f4c99c4f-6cnmx" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon" containerID="cri-o://22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c" gracePeriod=30 Oct 10 16:27:26 crc kubenswrapper[4788]: I1010 16:27:26.265587 4788 generic.go:334] "Generic (PLEG): container finished" podID="29c792f5-3274-4ff6-b87b-942398c3671d" containerID="22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c" exitCode=0 Oct 10 16:27:26 crc kubenswrapper[4788]: I1010 16:27:26.265655 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f4c99c4f-6cnmx" event={"ID":"29c792f5-3274-4ff6-b87b-942398c3671d","Type":"ContainerDied","Data":"22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c"} Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.278311 4788 generic.go:334] "Generic (PLEG): container finished" podID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerID="815417e3409a3ffb3d8bd21f037b20c1cbc6aa030eb6bad68641d161f1a05ae4" exitCode=137 Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.278763 4788 generic.go:334] "Generic (PLEG): container finished" podID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerID="ae9c05815d38508e24dbac6c4f6c1270ac67bf7cf118b5523ebe04760f389751" exitCode=137 Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.278398 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb58f66f-8p4t2" event={"ID":"86ca1055-1050-4d17-8de6-e74a3d72eafc","Type":"ContainerDied","Data":"815417e3409a3ffb3d8bd21f037b20c1cbc6aa030eb6bad68641d161f1a05ae4"} Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.278804 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb58f66f-8p4t2" event={"ID":"86ca1055-1050-4d17-8de6-e74a3d72eafc","Type":"ContainerDied","Data":"ae9c05815d38508e24dbac6c4f6c1270ac67bf7cf118b5523ebe04760f389751"} Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.568436 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.585998 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-config-data\") pod \"86ca1055-1050-4d17-8de6-e74a3d72eafc\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.586293 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86ca1055-1050-4d17-8de6-e74a3d72eafc-logs\") pod \"86ca1055-1050-4d17-8de6-e74a3d72eafc\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.586323 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-scripts\") pod \"86ca1055-1050-4d17-8de6-e74a3d72eafc\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.586406 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlvpz\" (UniqueName: \"kubernetes.io/projected/86ca1055-1050-4d17-8de6-e74a3d72eafc-kube-api-access-wlvpz\") pod \"86ca1055-1050-4d17-8de6-e74a3d72eafc\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.586459 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/86ca1055-1050-4d17-8de6-e74a3d72eafc-horizon-secret-key\") pod \"86ca1055-1050-4d17-8de6-e74a3d72eafc\" (UID: \"86ca1055-1050-4d17-8de6-e74a3d72eafc\") " Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.586974 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86ca1055-1050-4d17-8de6-e74a3d72eafc-logs" (OuterVolumeSpecName: "logs") pod "86ca1055-1050-4d17-8de6-e74a3d72eafc" (UID: "86ca1055-1050-4d17-8de6-e74a3d72eafc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.587477 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86ca1055-1050-4d17-8de6-e74a3d72eafc-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.605458 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86ca1055-1050-4d17-8de6-e74a3d72eafc-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "86ca1055-1050-4d17-8de6-e74a3d72eafc" (UID: "86ca1055-1050-4d17-8de6-e74a3d72eafc"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.618339 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ca1055-1050-4d17-8de6-e74a3d72eafc-kube-api-access-wlvpz" (OuterVolumeSpecName: "kube-api-access-wlvpz") pod "86ca1055-1050-4d17-8de6-e74a3d72eafc" (UID: "86ca1055-1050-4d17-8de6-e74a3d72eafc"). InnerVolumeSpecName "kube-api-access-wlvpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.634744 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-scripts" (OuterVolumeSpecName: "scripts") pod "86ca1055-1050-4d17-8de6-e74a3d72eafc" (UID: "86ca1055-1050-4d17-8de6-e74a3d72eafc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.665834 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-config-data" (OuterVolumeSpecName: "config-data") pod "86ca1055-1050-4d17-8de6-e74a3d72eafc" (UID: "86ca1055-1050-4d17-8de6-e74a3d72eafc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.689466 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.689501 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/86ca1055-1050-4d17-8de6-e74a3d72eafc-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.689515 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlvpz\" (UniqueName: \"kubernetes.io/projected/86ca1055-1050-4d17-8de6-e74a3d72eafc-kube-api-access-wlvpz\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:27 crc kubenswrapper[4788]: I1010 16:27:27.689526 4788 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/86ca1055-1050-4d17-8de6-e74a3d72eafc-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:28 crc kubenswrapper[4788]: I1010 16:27:28.291668 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67cb58f66f-8p4t2" event={"ID":"86ca1055-1050-4d17-8de6-e74a3d72eafc","Type":"ContainerDied","Data":"2266d455cf89c2810060c440b14406aba5aa81f8dd18e1412c9e348eadc0d38b"} Oct 10 16:27:28 crc kubenswrapper[4788]: I1010 16:27:28.291714 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67cb58f66f-8p4t2" Oct 10 16:27:28 crc kubenswrapper[4788]: I1010 16:27:28.291748 4788 scope.go:117] "RemoveContainer" containerID="815417e3409a3ffb3d8bd21f037b20c1cbc6aa030eb6bad68641d161f1a05ae4" Oct 10 16:27:28 crc kubenswrapper[4788]: I1010 16:27:28.318361 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67cb58f66f-8p4t2"] Oct 10 16:27:28 crc kubenswrapper[4788]: I1010 16:27:28.326342 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67cb58f66f-8p4t2"] Oct 10 16:27:28 crc kubenswrapper[4788]: I1010 16:27:28.445480 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67f4c99c4f-6cnmx" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 10 16:27:28 crc kubenswrapper[4788]: I1010 16:27:28.496534 4788 scope.go:117] "RemoveContainer" containerID="ae9c05815d38508e24dbac6c4f6c1270ac67bf7cf118b5523ebe04760f389751" Oct 10 16:27:30 crc kubenswrapper[4788]: I1010 16:27:30.253303 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" path="/var/lib/kubelet/pods/86ca1055-1050-4d17-8de6-e74a3d72eafc/volumes" Oct 10 16:27:30 crc kubenswrapper[4788]: I1010 16:27:30.975412 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-57bbb94c77-drsds"] Oct 10 16:27:30 crc kubenswrapper[4788]: E1010 16:27:30.975844 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerName="horizon-log" Oct 10 16:27:30 crc kubenswrapper[4788]: I1010 16:27:30.975858 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerName="horizon-log" Oct 10 16:27:30 crc kubenswrapper[4788]: E1010 16:27:30.975903 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerName="horizon" Oct 10 16:27:30 crc kubenswrapper[4788]: I1010 16:27:30.975911 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerName="horizon" Oct 10 16:27:30 crc kubenswrapper[4788]: I1010 16:27:30.976085 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerName="horizon" Oct 10 16:27:30 crc kubenswrapper[4788]: I1010 16:27:30.976104 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ca1055-1050-4d17-8de6-e74a3d72eafc" containerName="horizon-log" Oct 10 16:27:30 crc kubenswrapper[4788]: I1010 16:27:30.977247 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:30 crc kubenswrapper[4788]: I1010 16:27:30.989773 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57bbb94c77-drsds"] Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.078910 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38bff86e-3dbc-45d4-b48e-236d63f80e30-scripts\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.078984 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p28fj\" (UniqueName: \"kubernetes.io/projected/38bff86e-3dbc-45d4-b48e-236d63f80e30-kube-api-access-p28fj\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.079031 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38bff86e-3dbc-45d4-b48e-236d63f80e30-config-data\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.079047 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38bff86e-3dbc-45d4-b48e-236d63f80e30-logs\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.079116 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/38bff86e-3dbc-45d4-b48e-236d63f80e30-horizon-secret-key\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.181296 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38bff86e-3dbc-45d4-b48e-236d63f80e30-config-data\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.181339 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38bff86e-3dbc-45d4-b48e-236d63f80e30-logs\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.181416 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/38bff86e-3dbc-45d4-b48e-236d63f80e30-horizon-secret-key\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.181511 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38bff86e-3dbc-45d4-b48e-236d63f80e30-scripts\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.181528 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p28fj\" (UniqueName: \"kubernetes.io/projected/38bff86e-3dbc-45d4-b48e-236d63f80e30-kube-api-access-p28fj\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.182053 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38bff86e-3dbc-45d4-b48e-236d63f80e30-logs\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.182483 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38bff86e-3dbc-45d4-b48e-236d63f80e30-config-data\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.182909 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38bff86e-3dbc-45d4-b48e-236d63f80e30-scripts\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.189568 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/38bff86e-3dbc-45d4-b48e-236d63f80e30-horizon-secret-key\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.202657 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p28fj\" (UniqueName: \"kubernetes.io/projected/38bff86e-3dbc-45d4-b48e-236d63f80e30-kube-api-access-p28fj\") pod \"horizon-57bbb94c77-drsds\" (UID: \"38bff86e-3dbc-45d4-b48e-236d63f80e30\") " pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:31 crc kubenswrapper[4788]: I1010 16:27:31.312995 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.008095 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57bbb94c77-drsds"] Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.352709 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bbb94c77-drsds" event={"ID":"38bff86e-3dbc-45d4-b48e-236d63f80e30","Type":"ContainerStarted","Data":"65f13b9cf6f7811e37ca04124830edc87db38fbf73448d704458184cac9467bc"} Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.353072 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bbb94c77-drsds" event={"ID":"38bff86e-3dbc-45d4-b48e-236d63f80e30","Type":"ContainerStarted","Data":"31ec6ad07a39c841c9f91361a11260bead499e6bfbc3918c8ade7f5c015eabc2"} Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.741111 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-wbld9"] Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.742510 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-wbld9" Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.755290 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-wbld9"] Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.830510 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wj4s\" (UniqueName: \"kubernetes.io/projected/6066b1c7-e0b7-48a8-abd5-2d402169d6f7-kube-api-access-5wj4s\") pod \"heat-db-create-wbld9\" (UID: \"6066b1c7-e0b7-48a8-abd5-2d402169d6f7\") " pod="openstack/heat-db-create-wbld9" Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.932386 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wj4s\" (UniqueName: \"kubernetes.io/projected/6066b1c7-e0b7-48a8-abd5-2d402169d6f7-kube-api-access-5wj4s\") pod \"heat-db-create-wbld9\" (UID: \"6066b1c7-e0b7-48a8-abd5-2d402169d6f7\") " pod="openstack/heat-db-create-wbld9" Oct 10 16:27:32 crc kubenswrapper[4788]: I1010 16:27:32.953475 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wj4s\" (UniqueName: \"kubernetes.io/projected/6066b1c7-e0b7-48a8-abd5-2d402169d6f7-kube-api-access-5wj4s\") pod \"heat-db-create-wbld9\" (UID: \"6066b1c7-e0b7-48a8-abd5-2d402169d6f7\") " pod="openstack/heat-db-create-wbld9" Oct 10 16:27:33 crc kubenswrapper[4788]: I1010 16:27:33.061011 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-wbld9" Oct 10 16:27:33 crc kubenswrapper[4788]: I1010 16:27:33.371484 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bbb94c77-drsds" event={"ID":"38bff86e-3dbc-45d4-b48e-236d63f80e30","Type":"ContainerStarted","Data":"7be6ba7e93cff18412602e9e1961a7d697386a305d2726fa6f03812453cb16ff"} Oct 10 16:27:33 crc kubenswrapper[4788]: I1010 16:27:33.391933 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-57bbb94c77-drsds" podStartSLOduration=3.391913913 podStartE2EDuration="3.391913913s" podCreationTimestamp="2025-10-10 16:27:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:27:33.391606115 +0000 UTC m=+6155.841321673" watchObservedRunningTime="2025-10-10 16:27:33.391913913 +0000 UTC m=+6155.841629461" Oct 10 16:27:33 crc kubenswrapper[4788]: I1010 16:27:33.596905 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-wbld9"] Oct 10 16:27:34 crc kubenswrapper[4788]: I1010 16:27:34.384318 4788 generic.go:334] "Generic (PLEG): container finished" podID="6066b1c7-e0b7-48a8-abd5-2d402169d6f7" containerID="17c9642df5822b0e0247784a9c5729443b55c55d19cacf2ae42cae42db60ec9b" exitCode=0 Oct 10 16:27:34 crc kubenswrapper[4788]: I1010 16:27:34.384403 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-wbld9" event={"ID":"6066b1c7-e0b7-48a8-abd5-2d402169d6f7","Type":"ContainerDied","Data":"17c9642df5822b0e0247784a9c5729443b55c55d19cacf2ae42cae42db60ec9b"} Oct 10 16:27:34 crc kubenswrapper[4788]: I1010 16:27:34.384892 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-wbld9" event={"ID":"6066b1c7-e0b7-48a8-abd5-2d402169d6f7","Type":"ContainerStarted","Data":"5d29480c3d3dce99c01d5f57d97c30ecaf6d8eb3d096fb49768e1581df156787"} Oct 10 16:27:35 crc kubenswrapper[4788]: I1010 16:27:35.857804 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-wbld9" Oct 10 16:27:35 crc kubenswrapper[4788]: I1010 16:27:35.905614 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wj4s\" (UniqueName: \"kubernetes.io/projected/6066b1c7-e0b7-48a8-abd5-2d402169d6f7-kube-api-access-5wj4s\") pod \"6066b1c7-e0b7-48a8-abd5-2d402169d6f7\" (UID: \"6066b1c7-e0b7-48a8-abd5-2d402169d6f7\") " Oct 10 16:27:35 crc kubenswrapper[4788]: I1010 16:27:35.915629 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6066b1c7-e0b7-48a8-abd5-2d402169d6f7-kube-api-access-5wj4s" (OuterVolumeSpecName: "kube-api-access-5wj4s") pod "6066b1c7-e0b7-48a8-abd5-2d402169d6f7" (UID: "6066b1c7-e0b7-48a8-abd5-2d402169d6f7"). InnerVolumeSpecName "kube-api-access-5wj4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:27:36 crc kubenswrapper[4788]: I1010 16:27:36.008966 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wj4s\" (UniqueName: \"kubernetes.io/projected/6066b1c7-e0b7-48a8-abd5-2d402169d6f7-kube-api-access-5wj4s\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:36 crc kubenswrapper[4788]: I1010 16:27:36.407248 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-wbld9" event={"ID":"6066b1c7-e0b7-48a8-abd5-2d402169d6f7","Type":"ContainerDied","Data":"5d29480c3d3dce99c01d5f57d97c30ecaf6d8eb3d096fb49768e1581df156787"} Oct 10 16:27:36 crc kubenswrapper[4788]: I1010 16:27:36.407665 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d29480c3d3dce99c01d5f57d97c30ecaf6d8eb3d096fb49768e1581df156787" Oct 10 16:27:36 crc kubenswrapper[4788]: I1010 16:27:36.407322 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-wbld9" Oct 10 16:27:38 crc kubenswrapper[4788]: I1010 16:27:38.446203 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67f4c99c4f-6cnmx" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 10 16:27:41 crc kubenswrapper[4788]: I1010 16:27:41.313100 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:41 crc kubenswrapper[4788]: I1010 16:27:41.314022 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:42 crc kubenswrapper[4788]: I1010 16:27:42.864831 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-a27e-account-create-6lzql"] Oct 10 16:27:42 crc kubenswrapper[4788]: E1010 16:27:42.865713 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6066b1c7-e0b7-48a8-abd5-2d402169d6f7" containerName="mariadb-database-create" Oct 10 16:27:42 crc kubenswrapper[4788]: I1010 16:27:42.865731 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6066b1c7-e0b7-48a8-abd5-2d402169d6f7" containerName="mariadb-database-create" Oct 10 16:27:42 crc kubenswrapper[4788]: I1010 16:27:42.865935 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6066b1c7-e0b7-48a8-abd5-2d402169d6f7" containerName="mariadb-database-create" Oct 10 16:27:42 crc kubenswrapper[4788]: I1010 16:27:42.866771 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a27e-account-create-6lzql" Oct 10 16:27:42 crc kubenswrapper[4788]: I1010 16:27:42.885909 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 10 16:27:42 crc kubenswrapper[4788]: I1010 16:27:42.919199 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a27e-account-create-6lzql"] Oct 10 16:27:42 crc kubenswrapper[4788]: I1010 16:27:42.954638 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49vr9\" (UniqueName: \"kubernetes.io/projected/93f69a84-4f21-4331-afdc-657f573a2039-kube-api-access-49vr9\") pod \"heat-a27e-account-create-6lzql\" (UID: \"93f69a84-4f21-4331-afdc-657f573a2039\") " pod="openstack/heat-a27e-account-create-6lzql" Oct 10 16:27:43 crc kubenswrapper[4788]: I1010 16:27:43.057935 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49vr9\" (UniqueName: \"kubernetes.io/projected/93f69a84-4f21-4331-afdc-657f573a2039-kube-api-access-49vr9\") pod \"heat-a27e-account-create-6lzql\" (UID: \"93f69a84-4f21-4331-afdc-657f573a2039\") " pod="openstack/heat-a27e-account-create-6lzql" Oct 10 16:27:43 crc kubenswrapper[4788]: I1010 16:27:43.103987 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49vr9\" (UniqueName: \"kubernetes.io/projected/93f69a84-4f21-4331-afdc-657f573a2039-kube-api-access-49vr9\") pod \"heat-a27e-account-create-6lzql\" (UID: \"93f69a84-4f21-4331-afdc-657f573a2039\") " pod="openstack/heat-a27e-account-create-6lzql" Oct 10 16:27:43 crc kubenswrapper[4788]: I1010 16:27:43.220084 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a27e-account-create-6lzql" Oct 10 16:27:43 crc kubenswrapper[4788]: I1010 16:27:43.728499 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-a27e-account-create-6lzql"] Oct 10 16:27:44 crc kubenswrapper[4788]: I1010 16:27:44.045695 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-2brsx"] Oct 10 16:27:44 crc kubenswrapper[4788]: I1010 16:27:44.056527 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-2brsx"] Oct 10 16:27:44 crc kubenswrapper[4788]: I1010 16:27:44.246415 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e49ffce-869d-4edc-ac30-80ce5040803c" path="/var/lib/kubelet/pods/0e49ffce-869d-4edc-ac30-80ce5040803c/volumes" Oct 10 16:27:44 crc kubenswrapper[4788]: I1010 16:27:44.521965 4788 generic.go:334] "Generic (PLEG): container finished" podID="93f69a84-4f21-4331-afdc-657f573a2039" containerID="13b813173dae0404d579256c1e5c19a9b9d8de0c3b91d7d04313fad2ee568443" exitCode=0 Oct 10 16:27:44 crc kubenswrapper[4788]: I1010 16:27:44.522017 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a27e-account-create-6lzql" event={"ID":"93f69a84-4f21-4331-afdc-657f573a2039","Type":"ContainerDied","Data":"13b813173dae0404d579256c1e5c19a9b9d8de0c3b91d7d04313fad2ee568443"} Oct 10 16:27:44 crc kubenswrapper[4788]: I1010 16:27:44.522050 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a27e-account-create-6lzql" event={"ID":"93f69a84-4f21-4331-afdc-657f573a2039","Type":"ContainerStarted","Data":"c7b954ccdb4c0b387f37e9bbc0fa6788dbdf853c127a62a9eb74262d6796ce51"} Oct 10 16:27:45 crc kubenswrapper[4788]: I1010 16:27:45.986163 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a27e-account-create-6lzql" Oct 10 16:27:46 crc kubenswrapper[4788]: I1010 16:27:46.134954 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49vr9\" (UniqueName: \"kubernetes.io/projected/93f69a84-4f21-4331-afdc-657f573a2039-kube-api-access-49vr9\") pod \"93f69a84-4f21-4331-afdc-657f573a2039\" (UID: \"93f69a84-4f21-4331-afdc-657f573a2039\") " Oct 10 16:27:46 crc kubenswrapper[4788]: I1010 16:27:46.142589 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93f69a84-4f21-4331-afdc-657f573a2039-kube-api-access-49vr9" (OuterVolumeSpecName: "kube-api-access-49vr9") pod "93f69a84-4f21-4331-afdc-657f573a2039" (UID: "93f69a84-4f21-4331-afdc-657f573a2039"). InnerVolumeSpecName "kube-api-access-49vr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:27:46 crc kubenswrapper[4788]: I1010 16:27:46.240379 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49vr9\" (UniqueName: \"kubernetes.io/projected/93f69a84-4f21-4331-afdc-657f573a2039-kube-api-access-49vr9\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:46 crc kubenswrapper[4788]: I1010 16:27:46.548254 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-a27e-account-create-6lzql" event={"ID":"93f69a84-4f21-4331-afdc-657f573a2039","Type":"ContainerDied","Data":"c7b954ccdb4c0b387f37e9bbc0fa6788dbdf853c127a62a9eb74262d6796ce51"} Oct 10 16:27:46 crc kubenswrapper[4788]: I1010 16:27:46.548301 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7b954ccdb4c0b387f37e9bbc0fa6788dbdf853c127a62a9eb74262d6796ce51" Oct 10 16:27:46 crc kubenswrapper[4788]: I1010 16:27:46.548358 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-a27e-account-create-6lzql" Oct 10 16:27:47 crc kubenswrapper[4788]: I1010 16:27:47.926993 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-mg4lr"] Oct 10 16:27:47 crc kubenswrapper[4788]: E1010 16:27:47.928019 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f69a84-4f21-4331-afdc-657f573a2039" containerName="mariadb-account-create" Oct 10 16:27:47 crc kubenswrapper[4788]: I1010 16:27:47.928041 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f69a84-4f21-4331-afdc-657f573a2039" containerName="mariadb-account-create" Oct 10 16:27:47 crc kubenswrapper[4788]: I1010 16:27:47.928572 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f69a84-4f21-4331-afdc-657f573a2039" containerName="mariadb-account-create" Oct 10 16:27:47 crc kubenswrapper[4788]: I1010 16:27:47.929759 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:47 crc kubenswrapper[4788]: I1010 16:27:47.934488 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-bbrmf" Oct 10 16:27:47 crc kubenswrapper[4788]: I1010 16:27:47.934950 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 10 16:27:47 crc kubenswrapper[4788]: I1010 16:27:47.939164 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-mg4lr"] Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.077967 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-combined-ca-bundle\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.078415 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-config-data\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.078586 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhjq5\" (UniqueName: \"kubernetes.io/projected/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-kube-api-access-fhjq5\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.180909 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-combined-ca-bundle\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.181438 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-config-data\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.181741 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhjq5\" (UniqueName: \"kubernetes.io/projected/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-kube-api-access-fhjq5\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.195348 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-combined-ca-bundle\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.195430 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-config-data\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.206231 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhjq5\" (UniqueName: \"kubernetes.io/projected/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-kube-api-access-fhjq5\") pod \"heat-db-sync-mg4lr\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.263698 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mg4lr" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.448608 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-67f4c99c4f-6cnmx" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.448739 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:27:48 crc kubenswrapper[4788]: I1010 16:27:48.955038 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-mg4lr"] Oct 10 16:27:48 crc kubenswrapper[4788]: W1010 16:27:48.966624 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod842a80ee_3ba8_48ed_9513_dd197f1e2cd8.slice/crio-796225fb0116213b927c688ca8c326d3dbf1e3e5e53e8f13d8e035c24add0373 WatchSource:0}: Error finding container 796225fb0116213b927c688ca8c326d3dbf1e3e5e53e8f13d8e035c24add0373: Status 404 returned error can't find the container with id 796225fb0116213b927c688ca8c326d3dbf1e3e5e53e8f13d8e035c24add0373 Oct 10 16:27:49 crc kubenswrapper[4788]: I1010 16:27:49.575530 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mg4lr" event={"ID":"842a80ee-3ba8-48ed-9513-dd197f1e2cd8","Type":"ContainerStarted","Data":"796225fb0116213b927c688ca8c326d3dbf1e3e5e53e8f13d8e035c24add0373"} Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.193909 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.219739 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.284110 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql725\" (UniqueName: \"kubernetes.io/projected/29c792f5-3274-4ff6-b87b-942398c3671d-kube-api-access-ql725\") pod \"29c792f5-3274-4ff6-b87b-942398c3671d\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.284174 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c792f5-3274-4ff6-b87b-942398c3671d-logs\") pod \"29c792f5-3274-4ff6-b87b-942398c3671d\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.284256 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-config-data\") pod \"29c792f5-3274-4ff6-b87b-942398c3671d\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.284300 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-scripts\") pod \"29c792f5-3274-4ff6-b87b-942398c3671d\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.284442 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/29c792f5-3274-4ff6-b87b-942398c3671d-horizon-secret-key\") pod \"29c792f5-3274-4ff6-b87b-942398c3671d\" (UID: \"29c792f5-3274-4ff6-b87b-942398c3671d\") " Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.284772 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29c792f5-3274-4ff6-b87b-942398c3671d-logs" (OuterVolumeSpecName: "logs") pod "29c792f5-3274-4ff6-b87b-942398c3671d" (UID: "29c792f5-3274-4ff6-b87b-942398c3671d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.285035 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c792f5-3274-4ff6-b87b-942398c3671d-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.290163 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c792f5-3274-4ff6-b87b-942398c3671d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "29c792f5-3274-4ff6-b87b-942398c3671d" (UID: "29c792f5-3274-4ff6-b87b-942398c3671d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.290393 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29c792f5-3274-4ff6-b87b-942398c3671d-kube-api-access-ql725" (OuterVolumeSpecName: "kube-api-access-ql725") pod "29c792f5-3274-4ff6-b87b-942398c3671d" (UID: "29c792f5-3274-4ff6-b87b-942398c3671d"). InnerVolumeSpecName "kube-api-access-ql725". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.310046 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-config-data" (OuterVolumeSpecName: "config-data") pod "29c792f5-3274-4ff6-b87b-942398c3671d" (UID: "29c792f5-3274-4ff6-b87b-942398c3671d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.321280 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-scripts" (OuterVolumeSpecName: "scripts") pod "29c792f5-3274-4ff6-b87b-942398c3671d" (UID: "29c792f5-3274-4ff6-b87b-942398c3671d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.388035 4788 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/29c792f5-3274-4ff6-b87b-942398c3671d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.388062 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql725\" (UniqueName: \"kubernetes.io/projected/29c792f5-3274-4ff6-b87b-942398c3671d-kube-api-access-ql725\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.388076 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.388086 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c792f5-3274-4ff6-b87b-942398c3671d-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.613076 4788 generic.go:334] "Generic (PLEG): container finished" podID="29c792f5-3274-4ff6-b87b-942398c3671d" containerID="13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2" exitCode=137 Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.613123 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f4c99c4f-6cnmx" event={"ID":"29c792f5-3274-4ff6-b87b-942398c3671d","Type":"ContainerDied","Data":"13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2"} Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.613162 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67f4c99c4f-6cnmx" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.613184 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67f4c99c4f-6cnmx" event={"ID":"29c792f5-3274-4ff6-b87b-942398c3671d","Type":"ContainerDied","Data":"7c6e7daf22b03423e67261f8b5f91b6ec094f3154a567e17735d5bd2f7ecb2c5"} Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.613208 4788 scope.go:117] "RemoveContainer" containerID="22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.655049 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67f4c99c4f-6cnmx"] Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.662110 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67f4c99c4f-6cnmx"] Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.825138 4788 scope.go:117] "RemoveContainer" containerID="13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.846678 4788 scope.go:117] "RemoveContainer" containerID="22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c" Oct 10 16:27:53 crc kubenswrapper[4788]: E1010 16:27:53.847279 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c\": container with ID starting with 22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c not found: ID does not exist" containerID="22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.847318 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c"} err="failed to get container status \"22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c\": rpc error: code = NotFound desc = could not find container \"22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c\": container with ID starting with 22bf9fa29e737eb98824600549470d4b70e4e21ef8e5d67051cb1486c358663c not found: ID does not exist" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.847358 4788 scope.go:117] "RemoveContainer" containerID="13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2" Oct 10 16:27:53 crc kubenswrapper[4788]: E1010 16:27:53.847795 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2\": container with ID starting with 13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2 not found: ID does not exist" containerID="13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2" Oct 10 16:27:53 crc kubenswrapper[4788]: I1010 16:27:53.847825 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2"} err="failed to get container status \"13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2\": rpc error: code = NotFound desc = could not find container \"13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2\": container with ID starting with 13324cdd4aa7bad16555a135917c609bc1f09f41ccbb87af906ab327db54acb2 not found: ID does not exist" Oct 10 16:27:54 crc kubenswrapper[4788]: I1010 16:27:54.044113 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c7cb-account-create-v5t9s"] Oct 10 16:27:54 crc kubenswrapper[4788]: I1010 16:27:54.055790 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c7cb-account-create-v5t9s"] Oct 10 16:27:54 crc kubenswrapper[4788]: I1010 16:27:54.258156 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" path="/var/lib/kubelet/pods/29c792f5-3274-4ff6-b87b-942398c3671d/volumes" Oct 10 16:27:54 crc kubenswrapper[4788]: I1010 16:27:54.259066 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2cd6400-29f0-41da-9c22-0443d88dca9d" path="/var/lib/kubelet/pods/f2cd6400-29f0-41da-9c22-0443d88dca9d/volumes" Oct 10 16:27:54 crc kubenswrapper[4788]: I1010 16:27:54.956682 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-57bbb94c77-drsds" Oct 10 16:27:55 crc kubenswrapper[4788]: I1010 16:27:55.022924 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65cbd96cf7-rw4r7"] Oct 10 16:27:55 crc kubenswrapper[4788]: I1010 16:27:55.023311 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65cbd96cf7-rw4r7" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon-log" containerID="cri-o://6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0" gracePeriod=30 Oct 10 16:27:55 crc kubenswrapper[4788]: I1010 16:27:55.023905 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65cbd96cf7-rw4r7" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon" containerID="cri-o://e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075" gracePeriod=30 Oct 10 16:27:58 crc kubenswrapper[4788]: I1010 16:27:58.673017 4788 generic.go:334] "Generic (PLEG): container finished" podID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerID="e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075" exitCode=0 Oct 10 16:27:58 crc kubenswrapper[4788]: I1010 16:27:58.673094 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65cbd96cf7-rw4r7" event={"ID":"a32e043f-56f6-4468-b2d5-ef00f3854626","Type":"ContainerDied","Data":"e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075"} Oct 10 16:27:59 crc kubenswrapper[4788]: I1010 16:27:59.209486 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65cbd96cf7-rw4r7" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Oct 10 16:27:59 crc kubenswrapper[4788]: I1010 16:27:59.695285 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mg4lr" event={"ID":"842a80ee-3ba8-48ed-9513-dd197f1e2cd8","Type":"ContainerStarted","Data":"71c05576423d8055d7b01451cc5e9c59286b40586cc0cdf9439a2466a5d48f67"} Oct 10 16:27:59 crc kubenswrapper[4788]: I1010 16:27:59.722402 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-mg4lr" podStartSLOduration=2.711080303 podStartE2EDuration="12.722378045s" podCreationTimestamp="2025-10-10 16:27:47 +0000 UTC" firstStartedPulling="2025-10-10 16:27:48.96936857 +0000 UTC m=+6171.419084118" lastFinishedPulling="2025-10-10 16:27:58.980666312 +0000 UTC m=+6181.430381860" observedRunningTime="2025-10-10 16:27:59.712591506 +0000 UTC m=+6182.162307054" watchObservedRunningTime="2025-10-10 16:27:59.722378045 +0000 UTC m=+6182.172093633" Oct 10 16:28:01 crc kubenswrapper[4788]: I1010 16:28:01.055837 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-d2bgn"] Oct 10 16:28:01 crc kubenswrapper[4788]: I1010 16:28:01.071377 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-d2bgn"] Oct 10 16:28:01 crc kubenswrapper[4788]: I1010 16:28:01.725293 4788 generic.go:334] "Generic (PLEG): container finished" podID="842a80ee-3ba8-48ed-9513-dd197f1e2cd8" containerID="71c05576423d8055d7b01451cc5e9c59286b40586cc0cdf9439a2466a5d48f67" exitCode=0 Oct 10 16:28:01 crc kubenswrapper[4788]: I1010 16:28:01.725342 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mg4lr" event={"ID":"842a80ee-3ba8-48ed-9513-dd197f1e2cd8","Type":"ContainerDied","Data":"71c05576423d8055d7b01451cc5e9c59286b40586cc0cdf9439a2466a5d48f67"} Oct 10 16:28:02 crc kubenswrapper[4788]: I1010 16:28:02.264289 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="495332ff-4fbc-44db-bb18-bd59715f8e4e" path="/var/lib/kubelet/pods/495332ff-4fbc-44db-bb18-bd59715f8e4e/volumes" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.108084 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mg4lr" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.284302 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhjq5\" (UniqueName: \"kubernetes.io/projected/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-kube-api-access-fhjq5\") pod \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.285225 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-combined-ca-bundle\") pod \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.285504 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-config-data\") pod \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\" (UID: \"842a80ee-3ba8-48ed-9513-dd197f1e2cd8\") " Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.291086 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-kube-api-access-fhjq5" (OuterVolumeSpecName: "kube-api-access-fhjq5") pod "842a80ee-3ba8-48ed-9513-dd197f1e2cd8" (UID: "842a80ee-3ba8-48ed-9513-dd197f1e2cd8"). InnerVolumeSpecName "kube-api-access-fhjq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.318969 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "842a80ee-3ba8-48ed-9513-dd197f1e2cd8" (UID: "842a80ee-3ba8-48ed-9513-dd197f1e2cd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.390483 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhjq5\" (UniqueName: \"kubernetes.io/projected/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-kube-api-access-fhjq5\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.390511 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.397290 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-config-data" (OuterVolumeSpecName: "config-data") pod "842a80ee-3ba8-48ed-9513-dd197f1e2cd8" (UID: "842a80ee-3ba8-48ed-9513-dd197f1e2cd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.492698 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/842a80ee-3ba8-48ed-9513-dd197f1e2cd8-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.750676 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mg4lr" event={"ID":"842a80ee-3ba8-48ed-9513-dd197f1e2cd8","Type":"ContainerDied","Data":"796225fb0116213b927c688ca8c326d3dbf1e3e5e53e8f13d8e035c24add0373"} Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.750749 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="796225fb0116213b927c688ca8c326d3dbf1e3e5e53e8f13d8e035c24add0373" Oct 10 16:28:03 crc kubenswrapper[4788]: I1010 16:28:03.750762 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mg4lr" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.929897 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7bcb6dbfb-zktx9"] Oct 10 16:28:04 crc kubenswrapper[4788]: E1010 16:28:04.930763 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon-log" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.930784 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon-log" Oct 10 16:28:04 crc kubenswrapper[4788]: E1010 16:28:04.930828 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="842a80ee-3ba8-48ed-9513-dd197f1e2cd8" containerName="heat-db-sync" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.930838 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="842a80ee-3ba8-48ed-9513-dd197f1e2cd8" containerName="heat-db-sync" Oct 10 16:28:04 crc kubenswrapper[4788]: E1010 16:28:04.930854 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.930864 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.931103 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.931129 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="842a80ee-3ba8-48ed-9513-dd197f1e2cd8" containerName="heat-db-sync" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.931164 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="29c792f5-3274-4ff6-b87b-942398c3671d" containerName="horizon-log" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.932080 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.939300 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-bbrmf" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.939297 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 10 16:28:04 crc kubenswrapper[4788]: I1010 16:28:04.939560 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.021743 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7bcb6dbfb-zktx9"] Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.024188 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-combined-ca-bundle\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.024259 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-config-data\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.024315 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n546\" (UniqueName: \"kubernetes.io/projected/5c439913-7ebb-4ec3-921f-a7951566ef4b-kube-api-access-5n546\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.024437 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-config-data-custom\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.054256 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-f7fd69bfb-kdvkm"] Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.056102 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.058113 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.076350 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f7fd69bfb-kdvkm"] Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.115659 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-dd78bbb65-x49zv"] Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.116934 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.119350 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126001 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-config-data-custom\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126055 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-combined-ca-bundle\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126089 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-config-data\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126130 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n546\" (UniqueName: \"kubernetes.io/projected/5c439913-7ebb-4ec3-921f-a7951566ef4b-kube-api-access-5n546\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126188 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-combined-ca-bundle\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126213 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-config-data-custom\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126235 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-combined-ca-bundle\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126255 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-config-data\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126276 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvcpc\" (UniqueName: \"kubernetes.io/projected/245f7946-b9af-40b8-b5f3-df13e6af95b8-kube-api-access-hvcpc\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126310 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-config-data\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126327 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-config-data-custom\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.126364 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxxv8\" (UniqueName: \"kubernetes.io/projected/1e8995e7-5e84-4faf-b2d7-90d333960392-kube-api-access-nxxv8\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.130841 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-dd78bbb65-x49zv"] Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.135845 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-combined-ca-bundle\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.140301 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-config-data\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.140892 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c439913-7ebb-4ec3-921f-a7951566ef4b-config-data-custom\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.150611 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n546\" (UniqueName: \"kubernetes.io/projected/5c439913-7ebb-4ec3-921f-a7951566ef4b-kube-api-access-5n546\") pod \"heat-engine-7bcb6dbfb-zktx9\" (UID: \"5c439913-7ebb-4ec3-921f-a7951566ef4b\") " pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.229501 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-combined-ca-bundle\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.229565 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-config-data-custom\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.229593 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-combined-ca-bundle\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.229619 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-config-data\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.229641 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvcpc\" (UniqueName: \"kubernetes.io/projected/245f7946-b9af-40b8-b5f3-df13e6af95b8-kube-api-access-hvcpc\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.229674 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-config-data\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.229716 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxxv8\" (UniqueName: \"kubernetes.io/projected/1e8995e7-5e84-4faf-b2d7-90d333960392-kube-api-access-nxxv8\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.229745 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-config-data-custom\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.234354 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-combined-ca-bundle\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.237060 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-config-data-custom\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.241296 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-config-data-custom\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.242897 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/245f7946-b9af-40b8-b5f3-df13e6af95b8-config-data\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.249014 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-combined-ca-bundle\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.252122 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxxv8\" (UniqueName: \"kubernetes.io/projected/1e8995e7-5e84-4faf-b2d7-90d333960392-kube-api-access-nxxv8\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.252999 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvcpc\" (UniqueName: \"kubernetes.io/projected/245f7946-b9af-40b8-b5f3-df13e6af95b8-kube-api-access-hvcpc\") pod \"heat-api-f7fd69bfb-kdvkm\" (UID: \"245f7946-b9af-40b8-b5f3-df13e6af95b8\") " pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.253422 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e8995e7-5e84-4faf-b2d7-90d333960392-config-data\") pod \"heat-cfnapi-dd78bbb65-x49zv\" (UID: \"1e8995e7-5e84-4faf-b2d7-90d333960392\") " pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.269607 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.374729 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.522943 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.796892 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7bcb6dbfb-zktx9"] Oct 10 16:28:05 crc kubenswrapper[4788]: W1010 16:28:05.911853 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod245f7946_b9af_40b8_b5f3_df13e6af95b8.slice/crio-698fd2364490c93875e851e02cfb3c8ed7a17d4f9e32497e929152545722c744 WatchSource:0}: Error finding container 698fd2364490c93875e851e02cfb3c8ed7a17d4f9e32497e929152545722c744: Status 404 returned error can't find the container with id 698fd2364490c93875e851e02cfb3c8ed7a17d4f9e32497e929152545722c744 Oct 10 16:28:05 crc kubenswrapper[4788]: I1010 16:28:05.911982 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f7fd69bfb-kdvkm"] Oct 10 16:28:06 crc kubenswrapper[4788]: W1010 16:28:06.091563 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e8995e7_5e84_4faf_b2d7_90d333960392.slice/crio-d39616c5a12a5835210127d6d4d4959d633f7567a26df4ff68b91fb8c4cef9b2 WatchSource:0}: Error finding container d39616c5a12a5835210127d6d4d4959d633f7567a26df4ff68b91fb8c4cef9b2: Status 404 returned error can't find the container with id d39616c5a12a5835210127d6d4d4959d633f7567a26df4ff68b91fb8c4cef9b2 Oct 10 16:28:06 crc kubenswrapper[4788]: I1010 16:28:06.092593 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-dd78bbb65-x49zv"] Oct 10 16:28:06 crc kubenswrapper[4788]: I1010 16:28:06.782692 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-dd78bbb65-x49zv" event={"ID":"1e8995e7-5e84-4faf-b2d7-90d333960392","Type":"ContainerStarted","Data":"d39616c5a12a5835210127d6d4d4959d633f7567a26df4ff68b91fb8c4cef9b2"} Oct 10 16:28:06 crc kubenswrapper[4788]: I1010 16:28:06.785458 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7bcb6dbfb-zktx9" event={"ID":"5c439913-7ebb-4ec3-921f-a7951566ef4b","Type":"ContainerStarted","Data":"a06a34c7495040085e81e80d61a8af5372adeb281102512fde8f217a1496bf8d"} Oct 10 16:28:06 crc kubenswrapper[4788]: I1010 16:28:06.785566 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7bcb6dbfb-zktx9" event={"ID":"5c439913-7ebb-4ec3-921f-a7951566ef4b","Type":"ContainerStarted","Data":"c6f1dad019573af1892c8d9ec2db14290cf0f9d63119347599ecb78f1bb714f5"} Oct 10 16:28:06 crc kubenswrapper[4788]: I1010 16:28:06.785603 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:06 crc kubenswrapper[4788]: I1010 16:28:06.787084 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f7fd69bfb-kdvkm" event={"ID":"245f7946-b9af-40b8-b5f3-df13e6af95b8","Type":"ContainerStarted","Data":"698fd2364490c93875e851e02cfb3c8ed7a17d4f9e32497e929152545722c744"} Oct 10 16:28:06 crc kubenswrapper[4788]: I1010 16:28:06.803863 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7bcb6dbfb-zktx9" podStartSLOduration=2.803842295 podStartE2EDuration="2.803842295s" podCreationTimestamp="2025-10-10 16:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:28:06.801026391 +0000 UTC m=+6189.250741939" watchObservedRunningTime="2025-10-10 16:28:06.803842295 +0000 UTC m=+6189.253557843" Oct 10 16:28:09 crc kubenswrapper[4788]: I1010 16:28:09.209975 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65cbd96cf7-rw4r7" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Oct 10 16:28:09 crc kubenswrapper[4788]: I1010 16:28:09.829921 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f7fd69bfb-kdvkm" event={"ID":"245f7946-b9af-40b8-b5f3-df13e6af95b8","Type":"ContainerStarted","Data":"f1b19cc90325f6275880a15c99274d2db94ed4adfcff2f86cfad9e6294a3e69a"} Oct 10 16:28:09 crc kubenswrapper[4788]: I1010 16:28:09.830686 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:09 crc kubenswrapper[4788]: I1010 16:28:09.857313 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-f7fd69bfb-kdvkm" podStartSLOduration=2.121136532 podStartE2EDuration="4.857295511s" podCreationTimestamp="2025-10-10 16:28:05 +0000 UTC" firstStartedPulling="2025-10-10 16:28:05.91355555 +0000 UTC m=+6188.363271098" lastFinishedPulling="2025-10-10 16:28:08.649714529 +0000 UTC m=+6191.099430077" observedRunningTime="2025-10-10 16:28:09.851803706 +0000 UTC m=+6192.301519264" watchObservedRunningTime="2025-10-10 16:28:09.857295511 +0000 UTC m=+6192.307011059" Oct 10 16:28:14 crc kubenswrapper[4788]: I1010 16:28:14.503616 4788 scope.go:117] "RemoveContainer" containerID="f7badbc58fd5f032a6b5505500f9bcdf2769a462dfe938366f55383bd7f1877f" Oct 10 16:28:14 crc kubenswrapper[4788]: I1010 16:28:14.547734 4788 scope.go:117] "RemoveContainer" containerID="c891a5dfbf246674ff54c6a3fb9e32fd9ad162df65a95aeb0f639bfed9af6f87" Oct 10 16:28:14 crc kubenswrapper[4788]: I1010 16:28:14.606841 4788 scope.go:117] "RemoveContainer" containerID="516a629af29b70aa5c5fb6a8ad3bb4fa3a67c404433276b525663a3b71197d1f" Oct 10 16:28:15 crc kubenswrapper[4788]: I1010 16:28:15.298183 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7bcb6dbfb-zktx9" Oct 10 16:28:16 crc kubenswrapper[4788]: I1010 16:28:16.822513 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-f7fd69bfb-kdvkm" Oct 10 16:28:19 crc kubenswrapper[4788]: I1010 16:28:19.209833 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65cbd96cf7-rw4r7" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Oct 10 16:28:19 crc kubenswrapper[4788]: I1010 16:28:19.210922 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.572644 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.599793 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8"] Oct 10 16:28:25 crc kubenswrapper[4788]: E1010 16:28:25.600247 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon-log" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.600265 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon-log" Oct 10 16:28:25 crc kubenswrapper[4788]: E1010 16:28:25.600279 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.600286 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.600483 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.600514 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerName="horizon-log" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.601943 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.604669 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.614495 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8"] Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.673164 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-scripts\") pod \"a32e043f-56f6-4468-b2d5-ef00f3854626\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.673270 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j52z\" (UniqueName: \"kubernetes.io/projected/a32e043f-56f6-4468-b2d5-ef00f3854626-kube-api-access-9j52z\") pod \"a32e043f-56f6-4468-b2d5-ef00f3854626\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.673317 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-config-data\") pod \"a32e043f-56f6-4468-b2d5-ef00f3854626\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.673490 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32e043f-56f6-4468-b2d5-ef00f3854626-logs\") pod \"a32e043f-56f6-4468-b2d5-ef00f3854626\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.673550 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a32e043f-56f6-4468-b2d5-ef00f3854626-horizon-secret-key\") pod \"a32e043f-56f6-4468-b2d5-ef00f3854626\" (UID: \"a32e043f-56f6-4468-b2d5-ef00f3854626\") " Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.673954 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.674047 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmb9x\" (UniqueName: \"kubernetes.io/projected/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-kube-api-access-vmb9x\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.674117 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.674309 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a32e043f-56f6-4468-b2d5-ef00f3854626-logs" (OuterVolumeSpecName: "logs") pod "a32e043f-56f6-4468-b2d5-ef00f3854626" (UID: "a32e043f-56f6-4468-b2d5-ef00f3854626"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.682230 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a32e043f-56f6-4468-b2d5-ef00f3854626-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a32e043f-56f6-4468-b2d5-ef00f3854626" (UID: "a32e043f-56f6-4468-b2d5-ef00f3854626"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.697461 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-config-data" (OuterVolumeSpecName: "config-data") pod "a32e043f-56f6-4468-b2d5-ef00f3854626" (UID: "a32e043f-56f6-4468-b2d5-ef00f3854626"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.705707 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-scripts" (OuterVolumeSpecName: "scripts") pod "a32e043f-56f6-4468-b2d5-ef00f3854626" (UID: "a32e043f-56f6-4468-b2d5-ef00f3854626"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.706598 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32e043f-56f6-4468-b2d5-ef00f3854626-kube-api-access-9j52z" (OuterVolumeSpecName: "kube-api-access-9j52z") pod "a32e043f-56f6-4468-b2d5-ef00f3854626" (UID: "a32e043f-56f6-4468-b2d5-ef00f3854626"). InnerVolumeSpecName "kube-api-access-9j52z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.776780 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.777511 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmb9x\" (UniqueName: \"kubernetes.io/projected/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-kube-api-access-vmb9x\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.777734 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.777746 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.778224 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.778261 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.778441 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j52z\" (UniqueName: \"kubernetes.io/projected/a32e043f-56f6-4468-b2d5-ef00f3854626-kube-api-access-9j52z\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.778519 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a32e043f-56f6-4468-b2d5-ef00f3854626-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.778585 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a32e043f-56f6-4468-b2d5-ef00f3854626-logs\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.778652 4788 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a32e043f-56f6-4468-b2d5-ef00f3854626-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.796888 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmb9x\" (UniqueName: \"kubernetes.io/projected/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-kube-api-access-vmb9x\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:25 crc kubenswrapper[4788]: I1010 16:28:25.922330 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.061676 4788 generic.go:334] "Generic (PLEG): container finished" podID="a32e043f-56f6-4468-b2d5-ef00f3854626" containerID="6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0" exitCode=137 Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.061983 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65cbd96cf7-rw4r7" event={"ID":"a32e043f-56f6-4468-b2d5-ef00f3854626","Type":"ContainerDied","Data":"6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0"} Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.062014 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65cbd96cf7-rw4r7" event={"ID":"a32e043f-56f6-4468-b2d5-ef00f3854626","Type":"ContainerDied","Data":"77011c0216e8ae9f5f141145a4fd20363e773ea4ccefaeea6380a26020dd5bfc"} Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.062030 4788 scope.go:117] "RemoveContainer" containerID="e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.062156 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65cbd96cf7-rw4r7" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.115936 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65cbd96cf7-rw4r7"] Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.129820 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-65cbd96cf7-rw4r7"] Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.257228 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32e043f-56f6-4468-b2d5-ef00f3854626" path="/var/lib/kubelet/pods/a32e043f-56f6-4468-b2d5-ef00f3854626/volumes" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.314201 4788 scope.go:117] "RemoveContainer" containerID="6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.334968 4788 scope.go:117] "RemoveContainer" containerID="e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075" Oct 10 16:28:26 crc kubenswrapper[4788]: E1010 16:28:26.335432 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075\": container with ID starting with e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075 not found: ID does not exist" containerID="e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.335467 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075"} err="failed to get container status \"e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075\": rpc error: code = NotFound desc = could not find container \"e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075\": container with ID starting with e7a53463b8157ebb9cb63869252f03ca9440461ce3357bf529aa7247bff5f075 not found: ID does not exist" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.335489 4788 scope.go:117] "RemoveContainer" containerID="6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0" Oct 10 16:28:26 crc kubenswrapper[4788]: E1010 16:28:26.338176 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0\": container with ID starting with 6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0 not found: ID does not exist" containerID="6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.338200 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0"} err="failed to get container status \"6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0\": rpc error: code = NotFound desc = could not find container \"6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0\": container with ID starting with 6cd5ed361dbb4fcebdd3a8a8ecfba0578e0080e707a84b28d62a1aa1f73c20f0 not found: ID does not exist" Oct 10 16:28:26 crc kubenswrapper[4788]: I1010 16:28:26.411174 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8"] Oct 10 16:28:27 crc kubenswrapper[4788]: I1010 16:28:27.081921 4788 generic.go:334] "Generic (PLEG): container finished" podID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerID="c3976177cd58bd367b78dab2b0be724ee751ce388d884cb753cb9f4840ce4a8e" exitCode=0 Oct 10 16:28:27 crc kubenswrapper[4788]: I1010 16:28:27.082253 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" event={"ID":"65ef9ce9-fa0d-4d4d-b31b-5e1392189319","Type":"ContainerDied","Data":"c3976177cd58bd367b78dab2b0be724ee751ce388d884cb753cb9f4840ce4a8e"} Oct 10 16:28:27 crc kubenswrapper[4788]: I1010 16:28:27.082288 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" event={"ID":"65ef9ce9-fa0d-4d4d-b31b-5e1392189319","Type":"ContainerStarted","Data":"5e3bcec84397fe9d2ee12ebda41724a8bd27fbec9f39fb2f5179a8eab557324f"} Oct 10 16:28:29 crc kubenswrapper[4788]: I1010 16:28:29.106725 4788 generic.go:334] "Generic (PLEG): container finished" podID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerID="5b83fd21f79d415d20f951be9058bd1972ed7e642b81e0c297d4a29bd166b36b" exitCode=0 Oct 10 16:28:29 crc kubenswrapper[4788]: I1010 16:28:29.106849 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" event={"ID":"65ef9ce9-fa0d-4d4d-b31b-5e1392189319","Type":"ContainerDied","Data":"5b83fd21f79d415d20f951be9058bd1972ed7e642b81e0c297d4a29bd166b36b"} Oct 10 16:28:30 crc kubenswrapper[4788]: I1010 16:28:30.120451 4788 generic.go:334] "Generic (PLEG): container finished" podID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerID="94d58c727c7aa0440d186d7406a59c972e65579744de8b577d34d0af48ae6ecc" exitCode=0 Oct 10 16:28:30 crc kubenswrapper[4788]: I1010 16:28:30.120617 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" event={"ID":"65ef9ce9-fa0d-4d4d-b31b-5e1392189319","Type":"ContainerDied","Data":"94d58c727c7aa0440d186d7406a59c972e65579744de8b577d34d0af48ae6ecc"} Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.625201 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.795595 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-util\") pod \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.795979 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmb9x\" (UniqueName: \"kubernetes.io/projected/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-kube-api-access-vmb9x\") pod \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.796021 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-bundle\") pod \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\" (UID: \"65ef9ce9-fa0d-4d4d-b31b-5e1392189319\") " Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.797891 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-bundle" (OuterVolumeSpecName: "bundle") pod "65ef9ce9-fa0d-4d4d-b31b-5e1392189319" (UID: "65ef9ce9-fa0d-4d4d-b31b-5e1392189319"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.802379 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-kube-api-access-vmb9x" (OuterVolumeSpecName: "kube-api-access-vmb9x") pod "65ef9ce9-fa0d-4d4d-b31b-5e1392189319" (UID: "65ef9ce9-fa0d-4d4d-b31b-5e1392189319"). InnerVolumeSpecName "kube-api-access-vmb9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.898521 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmb9x\" (UniqueName: \"kubernetes.io/projected/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-kube-api-access-vmb9x\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.898563 4788 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:31 crc kubenswrapper[4788]: I1010 16:28:31.916185 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-util" (OuterVolumeSpecName: "util") pod "65ef9ce9-fa0d-4d4d-b31b-5e1392189319" (UID: "65ef9ce9-fa0d-4d4d-b31b-5e1392189319"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:28:32 crc kubenswrapper[4788]: I1010 16:28:32.001137 4788 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/65ef9ce9-fa0d-4d4d-b31b-5e1392189319-util\") on node \"crc\" DevicePath \"\"" Oct 10 16:28:32 crc kubenswrapper[4788]: I1010 16:28:32.162521 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" event={"ID":"65ef9ce9-fa0d-4d4d-b31b-5e1392189319","Type":"ContainerDied","Data":"5e3bcec84397fe9d2ee12ebda41724a8bd27fbec9f39fb2f5179a8eab557324f"} Oct 10 16:28:32 crc kubenswrapper[4788]: I1010 16:28:32.162577 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e3bcec84397fe9d2ee12ebda41724a8bd27fbec9f39fb2f5179a8eab557324f" Oct 10 16:28:32 crc kubenswrapper[4788]: I1010 16:28:32.162675 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8" Oct 10 16:28:41 crc kubenswrapper[4788]: I1010 16:28:41.294868 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-dd78bbb65-x49zv" event={"ID":"1e8995e7-5e84-4faf-b2d7-90d333960392","Type":"ContainerStarted","Data":"f30c7391f2526938f9b7810e77ffa119cead47ef099d656f3bbfa54c677af3fc"} Oct 10 16:28:41 crc kubenswrapper[4788]: I1010 16:28:41.295821 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:41 crc kubenswrapper[4788]: I1010 16:28:41.340788 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-dd78bbb65-x49zv" podStartSLOduration=2.402742198 podStartE2EDuration="36.340767668s" podCreationTimestamp="2025-10-10 16:28:05 +0000 UTC" firstStartedPulling="2025-10-10 16:28:06.094292602 +0000 UTC m=+6188.544008150" lastFinishedPulling="2025-10-10 16:28:40.032318072 +0000 UTC m=+6222.482033620" observedRunningTime="2025-10-10 16:28:41.326119761 +0000 UTC m=+6223.775835309" watchObservedRunningTime="2025-10-10 16:28:41.340767668 +0000 UTC m=+6223.790483216" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.783366 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-78srq"] Oct 10 16:28:42 crc kubenswrapper[4788]: E1010 16:28:42.791828 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerName="pull" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.792353 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerName="pull" Oct 10 16:28:42 crc kubenswrapper[4788]: E1010 16:28:42.792382 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerName="util" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.792388 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerName="util" Oct 10 16:28:42 crc kubenswrapper[4788]: E1010 16:28:42.792401 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerName="extract" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.792407 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerName="extract" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.792646 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ef9ce9-fa0d-4d4d-b31b-5e1392189319" containerName="extract" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.793663 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-78srq" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.797058 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.797260 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.798438 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-mj486" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.802657 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-78srq"] Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.866646 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpmpt\" (UniqueName: \"kubernetes.io/projected/deb5ec8e-ffbc-4cb8-8404-3e0903aa4934-kube-api-access-cpmpt\") pod \"obo-prometheus-operator-7c8cf85677-78srq\" (UID: \"deb5ec8e-ffbc-4cb8-8404-3e0903aa4934\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-78srq" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.930528 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8"] Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.938137 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.940538 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-j9kkj" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.941818 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.955016 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8"] Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.968636 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpmpt\" (UniqueName: \"kubernetes.io/projected/deb5ec8e-ffbc-4cb8-8404-3e0903aa4934-kube-api-access-cpmpt\") pod \"obo-prometheus-operator-7c8cf85677-78srq\" (UID: \"deb5ec8e-ffbc-4cb8-8404-3e0903aa4934\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-78srq" Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.974748 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz"] Oct 10 16:28:42 crc kubenswrapper[4788]: I1010 16:28:42.976233 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.010105 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpmpt\" (UniqueName: \"kubernetes.io/projected/deb5ec8e-ffbc-4cb8-8404-3e0903aa4934-kube-api-access-cpmpt\") pod \"obo-prometheus-operator-7c8cf85677-78srq\" (UID: \"deb5ec8e-ffbc-4cb8-8404-3e0903aa4934\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-78srq" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.017429 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz"] Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.071370 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bf651f1-a6e1-458b-87f7-b5faec152462-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8\" (UID: \"3bf651f1-a6e1-458b-87f7-b5faec152462\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.071425 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/640304e5-e29f-4190-8615-b5e1f71d4f31-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz\" (UID: \"640304e5-e29f-4190-8615-b5e1f71d4f31\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.071554 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bf651f1-a6e1-458b-87f7-b5faec152462-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8\" (UID: \"3bf651f1-a6e1-458b-87f7-b5faec152462\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.071590 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/640304e5-e29f-4190-8615-b5e1f71d4f31-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz\" (UID: \"640304e5-e29f-4190-8615-b5e1f71d4f31\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.126625 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-78srq" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.130625 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qjr28"] Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.132292 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.134829 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.138567 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-vjtjp" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.157400 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qjr28"] Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.176075 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bf651f1-a6e1-458b-87f7-b5faec152462-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8\" (UID: \"3bf651f1-a6e1-458b-87f7-b5faec152462\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.176156 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/640304e5-e29f-4190-8615-b5e1f71d4f31-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz\" (UID: \"640304e5-e29f-4190-8615-b5e1f71d4f31\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.176237 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bf651f1-a6e1-458b-87f7-b5faec152462-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8\" (UID: \"3bf651f1-a6e1-458b-87f7-b5faec152462\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.176265 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/640304e5-e29f-4190-8615-b5e1f71d4f31-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz\" (UID: \"640304e5-e29f-4190-8615-b5e1f71d4f31\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.182163 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/640304e5-e29f-4190-8615-b5e1f71d4f31-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz\" (UID: \"640304e5-e29f-4190-8615-b5e1f71d4f31\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.183996 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/640304e5-e29f-4190-8615-b5e1f71d4f31-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz\" (UID: \"640304e5-e29f-4190-8615-b5e1f71d4f31\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.184905 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bf651f1-a6e1-458b-87f7-b5faec152462-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8\" (UID: \"3bf651f1-a6e1-458b-87f7-b5faec152462\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.208760 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bf651f1-a6e1-458b-87f7-b5faec152462-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8\" (UID: \"3bf651f1-a6e1-458b-87f7-b5faec152462\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.258688 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.277919 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j954p\" (UniqueName: \"kubernetes.io/projected/982e6e19-775c-45bf-86ff-47233ac227e2-kube-api-access-j954p\") pod \"observability-operator-cc5f78dfc-qjr28\" (UID: \"982e6e19-775c-45bf-86ff-47233ac227e2\") " pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.278125 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/982e6e19-775c-45bf-86ff-47233ac227e2-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qjr28\" (UID: \"982e6e19-775c-45bf-86ff-47233ac227e2\") " pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.300648 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.345443 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-p5kcr"] Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.346763 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.351459 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-28bjf" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.364885 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-p5kcr"] Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.386112 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j954p\" (UniqueName: \"kubernetes.io/projected/982e6e19-775c-45bf-86ff-47233ac227e2-kube-api-access-j954p\") pod \"observability-operator-cc5f78dfc-qjr28\" (UID: \"982e6e19-775c-45bf-86ff-47233ac227e2\") " pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.387002 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/982e6e19-775c-45bf-86ff-47233ac227e2-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qjr28\" (UID: \"982e6e19-775c-45bf-86ff-47233ac227e2\") " pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.412162 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/982e6e19-775c-45bf-86ff-47233ac227e2-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-qjr28\" (UID: \"982e6e19-775c-45bf-86ff-47233ac227e2\") " pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.423112 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j954p\" (UniqueName: \"kubernetes.io/projected/982e6e19-775c-45bf-86ff-47233ac227e2-kube-api-access-j954p\") pod \"observability-operator-cc5f78dfc-qjr28\" (UID: \"982e6e19-775c-45bf-86ff-47233ac227e2\") " pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.456462 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.490788 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/83e92919-6e65-42d4-9a5d-ce7c78831457-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-p5kcr\" (UID: \"83e92919-6e65-42d4-9a5d-ce7c78831457\") " pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.491353 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zchmz\" (UniqueName: \"kubernetes.io/projected/83e92919-6e65-42d4-9a5d-ce7c78831457-kube-api-access-zchmz\") pod \"perses-operator-54bc95c9fb-p5kcr\" (UID: \"83e92919-6e65-42d4-9a5d-ce7c78831457\") " pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.598669 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/83e92919-6e65-42d4-9a5d-ce7c78831457-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-p5kcr\" (UID: \"83e92919-6e65-42d4-9a5d-ce7c78831457\") " pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.598781 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zchmz\" (UniqueName: \"kubernetes.io/projected/83e92919-6e65-42d4-9a5d-ce7c78831457-kube-api-access-zchmz\") pod \"perses-operator-54bc95c9fb-p5kcr\" (UID: \"83e92919-6e65-42d4-9a5d-ce7c78831457\") " pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.602931 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/83e92919-6e65-42d4-9a5d-ce7c78831457-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-p5kcr\" (UID: \"83e92919-6e65-42d4-9a5d-ce7c78831457\") " pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.624871 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zchmz\" (UniqueName: \"kubernetes.io/projected/83e92919-6e65-42d4-9a5d-ce7c78831457-kube-api-access-zchmz\") pod \"perses-operator-54bc95c9fb-p5kcr\" (UID: \"83e92919-6e65-42d4-9a5d-ce7c78831457\") " pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:43 crc kubenswrapper[4788]: I1010 16:28:43.800695 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:44 crc kubenswrapper[4788]: I1010 16:28:44.020932 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8"] Oct 10 16:28:44 crc kubenswrapper[4788]: I1010 16:28:44.025050 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-78srq"] Oct 10 16:28:44 crc kubenswrapper[4788]: W1010 16:28:44.036320 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bf651f1_a6e1_458b_87f7_b5faec152462.slice/crio-8a08a592eb487ca91f5bd712700381ff31f1916916bb67aa43fafea8de7e88bf WatchSource:0}: Error finding container 8a08a592eb487ca91f5bd712700381ff31f1916916bb67aa43fafea8de7e88bf: Status 404 returned error can't find the container with id 8a08a592eb487ca91f5bd712700381ff31f1916916bb67aa43fafea8de7e88bf Oct 10 16:28:44 crc kubenswrapper[4788]: I1010 16:28:44.196665 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz"] Oct 10 16:28:44 crc kubenswrapper[4788]: I1010 16:28:44.369271 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-qjr28"] Oct 10 16:28:44 crc kubenswrapper[4788]: I1010 16:28:44.370766 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" event={"ID":"3bf651f1-a6e1-458b-87f7-b5faec152462","Type":"ContainerStarted","Data":"8a08a592eb487ca91f5bd712700381ff31f1916916bb67aa43fafea8de7e88bf"} Oct 10 16:28:44 crc kubenswrapper[4788]: I1010 16:28:44.372825 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-78srq" event={"ID":"deb5ec8e-ffbc-4cb8-8404-3e0903aa4934","Type":"ContainerStarted","Data":"e493cb9f4aecf599f863f95c644f4944b86c53919c92271c90ee2dd29187ddcf"} Oct 10 16:28:44 crc kubenswrapper[4788]: I1010 16:28:44.374512 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" event={"ID":"640304e5-e29f-4190-8615-b5e1f71d4f31","Type":"ContainerStarted","Data":"f84e4fec97e83ca78690e1866bd60abd9dd7a7b3435738fd28caf88d54ac6080"} Oct 10 16:28:44 crc kubenswrapper[4788]: W1010 16:28:44.383996 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod982e6e19_775c_45bf_86ff_47233ac227e2.slice/crio-85c01cb2549b518e185c99b26b2481e44be81ab197d3d3c6d397ba2fa2b8f61b WatchSource:0}: Error finding container 85c01cb2549b518e185c99b26b2481e44be81ab197d3d3c6d397ba2fa2b8f61b: Status 404 returned error can't find the container with id 85c01cb2549b518e185c99b26b2481e44be81ab197d3d3c6d397ba2fa2b8f61b Oct 10 16:28:44 crc kubenswrapper[4788]: I1010 16:28:44.497937 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-p5kcr"] Oct 10 16:28:44 crc kubenswrapper[4788]: W1010 16:28:44.509840 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83e92919_6e65_42d4_9a5d_ce7c78831457.slice/crio-db2ceabb381760e4893f2f238e5dd5401d70418c52f826581a9871509a0756ae WatchSource:0}: Error finding container db2ceabb381760e4893f2f238e5dd5401d70418c52f826581a9871509a0756ae: Status 404 returned error can't find the container with id db2ceabb381760e4893f2f238e5dd5401d70418c52f826581a9871509a0756ae Oct 10 16:28:45 crc kubenswrapper[4788]: I1010 16:28:45.406400 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" event={"ID":"83e92919-6e65-42d4-9a5d-ce7c78831457","Type":"ContainerStarted","Data":"db2ceabb381760e4893f2f238e5dd5401d70418c52f826581a9871509a0756ae"} Oct 10 16:28:45 crc kubenswrapper[4788]: I1010 16:28:45.410511 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" event={"ID":"982e6e19-775c-45bf-86ff-47233ac227e2","Type":"ContainerStarted","Data":"85c01cb2549b518e185c99b26b2481e44be81ab197d3d3c6d397ba2fa2b8f61b"} Oct 10 16:28:47 crc kubenswrapper[4788]: I1010 16:28:47.122686 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-dd78bbb65-x49zv" Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.521664 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" event={"ID":"3bf651f1-a6e1-458b-87f7-b5faec152462","Type":"ContainerStarted","Data":"9add8d2c89553cd7ebabf40ab84432ca94f7f0e7f639dcf41a81df93c9aca540"} Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.525098 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" event={"ID":"640304e5-e29f-4190-8615-b5e1f71d4f31","Type":"ContainerStarted","Data":"92ab0d96c0ac28065188bee07301c6b507687a5f9244b35c994e6a20788f7bcd"} Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.529005 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" event={"ID":"83e92919-6e65-42d4-9a5d-ce7c78831457","Type":"ContainerStarted","Data":"542fdfc730383bda598a2d71cd59b9afe4e28d964c875ef1b61b54a430de8d59"} Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.529130 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.532344 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" event={"ID":"982e6e19-775c-45bf-86ff-47233ac227e2","Type":"ContainerStarted","Data":"e37efb59121faa2fc8be0158f5efa801b27d1abf3e831028fb57512c8a0f9ec0"} Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.532605 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.547547 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8" podStartSLOduration=2.807882235 podStartE2EDuration="12.547529016s" podCreationTimestamp="2025-10-10 16:28:42 +0000 UTC" firstStartedPulling="2025-10-10 16:28:44.038345698 +0000 UTC m=+6226.488061246" lastFinishedPulling="2025-10-10 16:28:53.777992479 +0000 UTC m=+6236.227708027" observedRunningTime="2025-10-10 16:28:54.53859409 +0000 UTC m=+6236.988309638" watchObservedRunningTime="2025-10-10 16:28:54.547529016 +0000 UTC m=+6236.997244564" Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.612586 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.621536 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz" podStartSLOduration=3.048748893 podStartE2EDuration="12.621520809s" podCreationTimestamp="2025-10-10 16:28:42 +0000 UTC" firstStartedPulling="2025-10-10 16:28:44.205237964 +0000 UTC m=+6226.654953512" lastFinishedPulling="2025-10-10 16:28:53.77800988 +0000 UTC m=+6236.227725428" observedRunningTime="2025-10-10 16:28:54.617206256 +0000 UTC m=+6237.066921824" watchObservedRunningTime="2025-10-10 16:28:54.621520809 +0000 UTC m=+6237.071236347" Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.655191 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" podStartSLOduration=2.216987516 podStartE2EDuration="11.655170678s" podCreationTimestamp="2025-10-10 16:28:43 +0000 UTC" firstStartedPulling="2025-10-10 16:28:44.517668773 +0000 UTC m=+6226.967384321" lastFinishedPulling="2025-10-10 16:28:53.955851915 +0000 UTC m=+6236.405567483" observedRunningTime="2025-10-10 16:28:54.637289775 +0000 UTC m=+6237.087005323" watchObservedRunningTime="2025-10-10 16:28:54.655170678 +0000 UTC m=+6237.104886216" Oct 10 16:28:54 crc kubenswrapper[4788]: I1010 16:28:54.674639 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-qjr28" podStartSLOduration=2.045433166 podStartE2EDuration="11.674621341s" podCreationTimestamp="2025-10-10 16:28:43 +0000 UTC" firstStartedPulling="2025-10-10 16:28:44.387056484 +0000 UTC m=+6226.836772032" lastFinishedPulling="2025-10-10 16:28:54.016244649 +0000 UTC m=+6236.465960207" observedRunningTime="2025-10-10 16:28:54.670422111 +0000 UTC m=+6237.120137659" watchObservedRunningTime="2025-10-10 16:28:54.674621341 +0000 UTC m=+6237.124336889" Oct 10 16:28:55 crc kubenswrapper[4788]: I1010 16:28:55.543496 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-78srq" event={"ID":"deb5ec8e-ffbc-4cb8-8404-3e0903aa4934","Type":"ContainerStarted","Data":"7927da1075e920538ebe111367925f8ca9db3b7f378996676188e1b09d4958a2"} Oct 10 16:28:55 crc kubenswrapper[4788]: I1010 16:28:55.563790 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-78srq" podStartSLOduration=3.822356998 podStartE2EDuration="13.563771976s" podCreationTimestamp="2025-10-10 16:28:42 +0000 UTC" firstStartedPulling="2025-10-10 16:28:44.040895015 +0000 UTC m=+6226.490610563" lastFinishedPulling="2025-10-10 16:28:53.782309993 +0000 UTC m=+6236.232025541" observedRunningTime="2025-10-10 16:28:55.562131893 +0000 UTC m=+6238.011847441" watchObservedRunningTime="2025-10-10 16:28:55.563771976 +0000 UTC m=+6238.013487524" Oct 10 16:28:59 crc kubenswrapper[4788]: I1010 16:28:59.406042 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:28:59 crc kubenswrapper[4788]: I1010 16:28:59.406522 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.044758 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-nmcdb"] Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.052855 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jn9s4"] Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.060587 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4x45l"] Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.070329 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jn9s4"] Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.085221 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-nmcdb"] Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.088581 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4x45l"] Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.314009 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77a38426-b4b3-4576-99de-b430003962d5" path="/var/lib/kubelet/pods/77a38426-b4b3-4576-99de-b430003962d5/volumes" Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.314863 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f418231-e326-423e-948e-fa69f7a727e0" path="/var/lib/kubelet/pods/7f418231-e326-423e-948e-fa69f7a727e0/volumes" Oct 10 16:29:00 crc kubenswrapper[4788]: I1010 16:29:00.328218 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80be9ca4-09b7-4907-ac98-1423d0ffa711" path="/var/lib/kubelet/pods/80be9ca4-09b7-4907-ac98-1423d0ffa711/volumes" Oct 10 16:29:03 crc kubenswrapper[4788]: I1010 16:29:03.805824 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-p5kcr" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.216776 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.217312 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" containerName="openstackclient" containerID="cri-o://d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c" gracePeriod=2 Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.228986 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.291500 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 10 16:29:06 crc kubenswrapper[4788]: E1010 16:29:06.292195 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" containerName="openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.292216 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" containerName="openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.292418 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" containerName="openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.293117 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.318131 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.378964 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjxlt\" (UniqueName: \"kubernetes.io/projected/8133c683-728c-4758-9c20-6eddde277d95-kube-api-access-xjxlt\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.379046 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8133c683-728c-4758-9c20-6eddde277d95-openstack-config-secret\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.379104 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8133c683-728c-4758-9c20-6eddde277d95-openstack-config\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.480820 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjxlt\" (UniqueName: \"kubernetes.io/projected/8133c683-728c-4758-9c20-6eddde277d95-kube-api-access-xjxlt\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.480895 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8133c683-728c-4758-9c20-6eddde277d95-openstack-config-secret\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.480933 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8133c683-728c-4758-9c20-6eddde277d95-openstack-config\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.481831 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8133c683-728c-4758-9c20-6eddde277d95-openstack-config\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.487137 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8133c683-728c-4758-9c20-6eddde277d95-openstack-config-secret\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.527081 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.528397 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.530668 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-g2gpd" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.547788 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjxlt\" (UniqueName: \"kubernetes.io/projected/8133c683-728c-4758-9c20-6eddde277d95-kube-api-access-xjxlt\") pod \"openstackclient\" (UID: \"8133c683-728c-4758-9c20-6eddde277d95\") " pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.553563 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.628235 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.688717 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9hs4\" (UniqueName: \"kubernetes.io/projected/1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c-kube-api-access-w9hs4\") pod \"kube-state-metrics-0\" (UID: \"1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c\") " pod="openstack/kube-state-metrics-0" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.794446 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9hs4\" (UniqueName: \"kubernetes.io/projected/1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c-kube-api-access-w9hs4\") pod \"kube-state-metrics-0\" (UID: \"1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c\") " pod="openstack/kube-state-metrics-0" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.863908 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9hs4\" (UniqueName: \"kubernetes.io/projected/1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c-kube-api-access-w9hs4\") pod \"kube-state-metrics-0\" (UID: \"1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c\") " pod="openstack/kube-state-metrics-0" Oct 10 16:29:06 crc kubenswrapper[4788]: I1010 16:29:06.908544 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.211400 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.223939 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.226528 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-rmw7v" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.226734 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.226874 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.228029 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.231958 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.315362 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/766efb82-f5a8-44a0-b451-8ee5d7bdb508-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.315450 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/766efb82-f5a8-44a0-b451-8ee5d7bdb508-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.315493 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/766efb82-f5a8-44a0-b451-8ee5d7bdb508-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.315534 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/766efb82-f5a8-44a0-b451-8ee5d7bdb508-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.315582 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcgln\" (UniqueName: \"kubernetes.io/projected/766efb82-f5a8-44a0-b451-8ee5d7bdb508-kube-api-access-kcgln\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.315609 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/766efb82-f5a8-44a0-b451-8ee5d7bdb508-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.417932 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcgln\" (UniqueName: \"kubernetes.io/projected/766efb82-f5a8-44a0-b451-8ee5d7bdb508-kube-api-access-kcgln\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.417988 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/766efb82-f5a8-44a0-b451-8ee5d7bdb508-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.418043 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/766efb82-f5a8-44a0-b451-8ee5d7bdb508-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.418120 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/766efb82-f5a8-44a0-b451-8ee5d7bdb508-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.419013 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/766efb82-f5a8-44a0-b451-8ee5d7bdb508-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.419072 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/766efb82-f5a8-44a0-b451-8ee5d7bdb508-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.420603 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/766efb82-f5a8-44a0-b451-8ee5d7bdb508-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.428454 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/766efb82-f5a8-44a0-b451-8ee5d7bdb508-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.431550 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/766efb82-f5a8-44a0-b451-8ee5d7bdb508-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.436615 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/766efb82-f5a8-44a0-b451-8ee5d7bdb508-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.437481 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/766efb82-f5a8-44a0-b451-8ee5d7bdb508-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.457887 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcgln\" (UniqueName: \"kubernetes.io/projected/766efb82-f5a8-44a0-b451-8ee5d7bdb508-kube-api-access-kcgln\") pod \"alertmanager-metric-storage-0\" (UID: \"766efb82-f5a8-44a0-b451-8ee5d7bdb508\") " pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.589631 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.627206 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.653304 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.790049 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c","Type":"ContainerStarted","Data":"ef255fc96a98f9754095b48070f9de3142df1e2affd0e81c06039b24e64be87d"} Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.799312 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8133c683-728c-4758-9c20-6eddde277d95","Type":"ContainerStarted","Data":"7d7a529bef022bd7a7c64b581121c90ff0ba8e94ed0e3f4840c954ea386df2be"} Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.878936 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.889286 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.893427 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.894075 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.894103 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-brzpz" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.894536 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.903663 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.923685 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 10 16:29:07 crc kubenswrapper[4788]: I1010 16:29:07.946046 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.057008 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.057065 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.057099 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.057245 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c034424e-9dbf-4446-9c41-9870da4dc340\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c034424e-9dbf-4446-9c41-9870da4dc340\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.057279 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-config\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.057297 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.057326 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk28n\" (UniqueName: \"kubernetes.io/projected/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-kube-api-access-pk28n\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.057417 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.161462 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk28n\" (UniqueName: \"kubernetes.io/projected/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-kube-api-access-pk28n\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.161587 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.161640 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.161659 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.161679 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.161744 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c034424e-9dbf-4446-9c41-9870da4dc340\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c034424e-9dbf-4446-9c41-9870da4dc340\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.161769 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-config\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.161788 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.177939 4788 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.177975 4788 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c034424e-9dbf-4446-9c41-9870da4dc340\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c034424e-9dbf-4446-9c41-9870da4dc340\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6d1ee717f262705857b5d71141339aa4e11ff3d92c89b8664925e1f23cda833a/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.186240 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.189336 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.190628 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.191877 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-config\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.192567 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.193090 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.208356 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk28n\" (UniqueName: \"kubernetes.io/projected/2fd93ba6-6366-4f85-b5f3-cef3ed588a50-kube-api-access-pk28n\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.371347 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c034424e-9dbf-4446-9c41-9870da4dc340\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c034424e-9dbf-4446-9c41-9870da4dc340\") pod \"prometheus-metric-storage-0\" (UID: \"2fd93ba6-6366-4f85-b5f3-cef3ed588a50\") " pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.551741 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.593628 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.850309 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.854583 4788 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" podUID="8133c683-728c-4758-9c20-6eddde277d95" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.868868 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c","Type":"ContainerStarted","Data":"8852973ec7695d548b4f77727de444beacf431053acd320e38139eb7823dc012"} Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.869475 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.873983 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8133c683-728c-4758-9c20-6eddde277d95","Type":"ContainerStarted","Data":"9654ade5ad05429de9f8a6efbe598711180a2eaa1a7e748a42ef9d934bcad067"} Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.885276 4788 generic.go:334] "Generic (PLEG): container finished" podID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" containerID="d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c" exitCode=137 Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.885361 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.885607 4788 scope.go:117] "RemoveContainer" containerID="d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.896652 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.398878752 podStartE2EDuration="2.896634663s" podCreationTimestamp="2025-10-10 16:29:06 +0000 UTC" firstStartedPulling="2025-10-10 16:29:07.68446002 +0000 UTC m=+6250.134175568" lastFinishedPulling="2025-10-10 16:29:08.182215931 +0000 UTC m=+6250.631931479" observedRunningTime="2025-10-10 16:29:08.894180268 +0000 UTC m=+6251.343895816" watchObservedRunningTime="2025-10-10 16:29:08.896634663 +0000 UTC m=+6251.346350211" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.897670 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"766efb82-f5a8-44a0-b451-8ee5d7bdb508","Type":"ContainerStarted","Data":"bc623b1262091910c61878133daf9a2f77aa6d20be608964d3e35254172ee035"} Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.900636 4788 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" podUID="8133c683-728c-4758-9c20-6eddde277d95" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.930864 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.930843226 podStartE2EDuration="2.930843226s" podCreationTimestamp="2025-10-10 16:29:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:29:08.91962536 +0000 UTC m=+6251.369340898" watchObservedRunningTime="2025-10-10 16:29:08.930843226 +0000 UTC m=+6251.380558774" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.962486 4788 scope.go:117] "RemoveContainer" containerID="d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c" Oct 10 16:29:08 crc kubenswrapper[4788]: E1010 16:29:08.962949 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c\": container with ID starting with d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c not found: ID does not exist" containerID="d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.962975 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c"} err="failed to get container status \"d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c\": rpc error: code = NotFound desc = could not find container \"d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c\": container with ID starting with d5c21bb72a9ddda84da81ae87c36e430b08972c5ef5bb690311425734cc5b17c not found: ID does not exist" Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.991927 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wjjd\" (UniqueName: \"kubernetes.io/projected/7a53dce9-770e-4fbf-8635-f69cc25fafd6-kube-api-access-8wjjd\") pod \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.992219 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config\") pod \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " Oct 10 16:29:08 crc kubenswrapper[4788]: I1010 16:29:08.992271 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config-secret\") pod \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\" (UID: \"7a53dce9-770e-4fbf-8635-f69cc25fafd6\") " Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.002339 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a53dce9-770e-4fbf-8635-f69cc25fafd6-kube-api-access-8wjjd" (OuterVolumeSpecName: "kube-api-access-8wjjd") pod "7a53dce9-770e-4fbf-8635-f69cc25fafd6" (UID: "7a53dce9-770e-4fbf-8635-f69cc25fafd6"). InnerVolumeSpecName "kube-api-access-8wjjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.039653 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7a53dce9-770e-4fbf-8635-f69cc25fafd6" (UID: "7a53dce9-770e-4fbf-8635-f69cc25fafd6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.090211 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7a53dce9-770e-4fbf-8635-f69cc25fafd6" (UID: "7a53dce9-770e-4fbf-8635-f69cc25fafd6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.094671 4788 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.094703 4788 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7a53dce9-770e-4fbf-8635-f69cc25fafd6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.094715 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wjjd\" (UniqueName: \"kubernetes.io/projected/7a53dce9-770e-4fbf-8635-f69cc25fafd6-kube-api-access-8wjjd\") on node \"crc\" DevicePath \"\"" Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.215050 4788 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" podUID="8133c683-728c-4758-9c20-6eddde277d95" Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.351206 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 10 16:29:09 crc kubenswrapper[4788]: I1010 16:29:09.909079 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2fd93ba6-6366-4f85-b5f3-cef3ed588a50","Type":"ContainerStarted","Data":"613f9f13cf19f0f6d249d87c8c47aa6865a2053ee74223d0f8601c31fdd72812"} Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.061854 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c927-account-create-vwdhf"] Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.076567 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8011-account-create-hb7s5"] Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.092784 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0d75-account-create-kb2xh"] Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.101532 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8011-account-create-hb7s5"] Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.109834 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c927-account-create-vwdhf"] Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.120008 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0d75-account-create-kb2xh"] Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.266982 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1655641f-54ec-48c1-bbb6-21e801e6a54f" path="/var/lib/kubelet/pods/1655641f-54ec-48c1-bbb6-21e801e6a54f/volumes" Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.267998 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a53dce9-770e-4fbf-8635-f69cc25fafd6" path="/var/lib/kubelet/pods/7a53dce9-770e-4fbf-8635-f69cc25fafd6/volumes" Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.269294 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cf0be59-5a06-45c5-b913-d75c1e4096c0" path="/var/lib/kubelet/pods/7cf0be59-5a06-45c5-b913-d75c1e4096c0/volumes" Oct 10 16:29:10 crc kubenswrapper[4788]: I1010 16:29:10.269815 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7d44061-2a00-4a40-94c7-0c56e37d55c9" path="/var/lib/kubelet/pods/a7d44061-2a00-4a40-94c7-0c56e37d55c9/volumes" Oct 10 16:29:14 crc kubenswrapper[4788]: I1010 16:29:14.804610 4788 scope.go:117] "RemoveContainer" containerID="2ff29cbbeaef5875febfd11818203b781590c9a4d5ee1b53f60df32688f3cd5a" Oct 10 16:29:14 crc kubenswrapper[4788]: I1010 16:29:14.833411 4788 scope.go:117] "RemoveContainer" containerID="bd2e4d573e6fa9ca14b21d675411bc87317c1059fa7b0a0c9dec5b6c7ad38a68" Oct 10 16:29:15 crc kubenswrapper[4788]: I1010 16:29:15.221192 4788 scope.go:117] "RemoveContainer" containerID="09e234a86a10f7cbcf7659080fecb1c1a6137b12d15dd15790db7f8d42b12708" Oct 10 16:29:15 crc kubenswrapper[4788]: I1010 16:29:15.331390 4788 scope.go:117] "RemoveContainer" containerID="c4692339c259979db1616d69dbb17890301f186aa97725329c4e309b4e3fad24" Oct 10 16:29:15 crc kubenswrapper[4788]: I1010 16:29:15.353937 4788 scope.go:117] "RemoveContainer" containerID="99ae820df5735d54c6b0124c9827eac416b9507d6069078f1b4cd7d6a1b1e424" Oct 10 16:29:15 crc kubenswrapper[4788]: I1010 16:29:15.405310 4788 scope.go:117] "RemoveContainer" containerID="261022933c054799e06153249db6b73508f0a918af13e3afcbf4fdb0544deb67" Oct 10 16:29:15 crc kubenswrapper[4788]: I1010 16:29:15.968675 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"766efb82-f5a8-44a0-b451-8ee5d7bdb508","Type":"ContainerStarted","Data":"00e53207868e427bc6d601a7308bfb2c236e747b52ccd664a303ec1a30fbee94"} Oct 10 16:29:15 crc kubenswrapper[4788]: I1010 16:29:15.970298 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2fd93ba6-6366-4f85-b5f3-cef3ed588a50","Type":"ContainerStarted","Data":"d6ca690e163a941f2df51bdaebbd86bac8cf7d35349bb74d8a9e05fd01e18b65"} Oct 10 16:29:16 crc kubenswrapper[4788]: I1010 16:29:16.913712 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 10 16:29:20 crc kubenswrapper[4788]: I1010 16:29:20.066658 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w972c"] Oct 10 16:29:20 crc kubenswrapper[4788]: I1010 16:29:20.088402 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-w972c"] Oct 10 16:29:20 crc kubenswrapper[4788]: I1010 16:29:20.261271 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06b3f17f-fbea-4f87-b0eb-4fc601596d51" path="/var/lib/kubelet/pods/06b3f17f-fbea-4f87-b0eb-4fc601596d51/volumes" Oct 10 16:29:22 crc kubenswrapper[4788]: I1010 16:29:22.042649 4788 generic.go:334] "Generic (PLEG): container finished" podID="766efb82-f5a8-44a0-b451-8ee5d7bdb508" containerID="00e53207868e427bc6d601a7308bfb2c236e747b52ccd664a303ec1a30fbee94" exitCode=0 Oct 10 16:29:22 crc kubenswrapper[4788]: I1010 16:29:22.042754 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"766efb82-f5a8-44a0-b451-8ee5d7bdb508","Type":"ContainerDied","Data":"00e53207868e427bc6d601a7308bfb2c236e747b52ccd664a303ec1a30fbee94"} Oct 10 16:29:22 crc kubenswrapper[4788]: I1010 16:29:22.044676 4788 generic.go:334] "Generic (PLEG): container finished" podID="2fd93ba6-6366-4f85-b5f3-cef3ed588a50" containerID="d6ca690e163a941f2df51bdaebbd86bac8cf7d35349bb74d8a9e05fd01e18b65" exitCode=0 Oct 10 16:29:22 crc kubenswrapper[4788]: I1010 16:29:22.044709 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2fd93ba6-6366-4f85-b5f3-cef3ed588a50","Type":"ContainerDied","Data":"d6ca690e163a941f2df51bdaebbd86bac8cf7d35349bb74d8a9e05fd01e18b65"} Oct 10 16:29:25 crc kubenswrapper[4788]: I1010 16:29:25.076466 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"766efb82-f5a8-44a0-b451-8ee5d7bdb508","Type":"ContainerStarted","Data":"d25dfb0dc76826a5996cd21a9eb8e08b0c97156d416f8e607fc039ba8f315503"} Oct 10 16:29:29 crc kubenswrapper[4788]: I1010 16:29:29.127940 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"766efb82-f5a8-44a0-b451-8ee5d7bdb508","Type":"ContainerStarted","Data":"db93e911431ccf63d234bb347c15a5802caf17df4681c049746f01a05ed45430"} Oct 10 16:29:29 crc kubenswrapper[4788]: I1010 16:29:29.128586 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:29 crc kubenswrapper[4788]: I1010 16:29:29.129960 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2fd93ba6-6366-4f85-b5f3-cef3ed588a50","Type":"ContainerStarted","Data":"a7318d1e27e4826618bd65a14834a0ad1b093245586e3c14f5201986ff9d95ed"} Oct 10 16:29:29 crc kubenswrapper[4788]: I1010 16:29:29.132362 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 10 16:29:29 crc kubenswrapper[4788]: I1010 16:29:29.154713 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.05993672 podStartE2EDuration="22.154694675s" podCreationTimestamp="2025-10-10 16:29:07 +0000 UTC" firstStartedPulling="2025-10-10 16:29:08.67834338 +0000 UTC m=+6251.128058928" lastFinishedPulling="2025-10-10 16:29:24.773101335 +0000 UTC m=+6267.222816883" observedRunningTime="2025-10-10 16:29:29.14920751 +0000 UTC m=+6271.598923058" watchObservedRunningTime="2025-10-10 16:29:29.154694675 +0000 UTC m=+6271.604410213" Oct 10 16:29:29 crc kubenswrapper[4788]: I1010 16:29:29.406881 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:29:29 crc kubenswrapper[4788]: I1010 16:29:29.407253 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:29:32 crc kubenswrapper[4788]: I1010 16:29:32.166959 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2fd93ba6-6366-4f85-b5f3-cef3ed588a50","Type":"ContainerStarted","Data":"018de4e542417fbf67cf90f3daa1ade6d4f606edcfac4978d21402297840d9aa"} Oct 10 16:29:34 crc kubenswrapper[4788]: I1010 16:29:34.036754 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2pdgt"] Oct 10 16:29:34 crc kubenswrapper[4788]: I1010 16:29:34.048287 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2pdgt"] Oct 10 16:29:34 crc kubenswrapper[4788]: I1010 16:29:34.245389 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1cff6f4-b4e1-420d-9723-25d83b9ee079" path="/var/lib/kubelet/pods/a1cff6f4-b4e1-420d-9723-25d83b9ee079/volumes" Oct 10 16:29:35 crc kubenswrapper[4788]: I1010 16:29:35.026936 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-q4s2n"] Oct 10 16:29:35 crc kubenswrapper[4788]: I1010 16:29:35.039452 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-q4s2n"] Oct 10 16:29:36 crc kubenswrapper[4788]: I1010 16:29:36.214661 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2fd93ba6-6366-4f85-b5f3-cef3ed588a50","Type":"ContainerStarted","Data":"46e192cdf1c40f6b736bcf91f733570853fb000c16da8893c56ce1beb5aae9a0"} Oct 10 16:29:36 crc kubenswrapper[4788]: I1010 16:29:36.241829 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=4.511493939 podStartE2EDuration="30.241811896s" podCreationTimestamp="2025-10-10 16:29:06 +0000 UTC" firstStartedPulling="2025-10-10 16:29:09.373883654 +0000 UTC m=+6251.823599202" lastFinishedPulling="2025-10-10 16:29:35.104201611 +0000 UTC m=+6277.553917159" observedRunningTime="2025-10-10 16:29:36.238026176 +0000 UTC m=+6278.687741724" watchObservedRunningTime="2025-10-10 16:29:36.241811896 +0000 UTC m=+6278.691527454" Oct 10 16:29:36 crc kubenswrapper[4788]: I1010 16:29:36.248602 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d29917ed-6776-4a18-8b81-e1034a2f68f4" path="/var/lib/kubelet/pods/d29917ed-6776-4a18-8b81-e1034a2f68f4/volumes" Oct 10 16:29:38 crc kubenswrapper[4788]: I1010 16:29:38.552165 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:38 crc kubenswrapper[4788]: I1010 16:29:38.552454 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:38 crc kubenswrapper[4788]: I1010 16:29:38.555387 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:39 crc kubenswrapper[4788]: I1010 16:29:39.263421 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.509515 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.513390 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.520795 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.521120 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.535374 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.630150 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-config-data\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.630219 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-scripts\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.630332 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnpk6\" (UniqueName: \"kubernetes.io/projected/dfadac38-b67f-4652-a05e-1f9ee873ab5b-kube-api-access-bnpk6\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.630385 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-run-httpd\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.630464 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.630562 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.630848 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-log-httpd\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.732803 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.732936 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-log-httpd\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.732959 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-config-data\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.732987 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-scripts\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.733025 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnpk6\" (UniqueName: \"kubernetes.io/projected/dfadac38-b67f-4652-a05e-1f9ee873ab5b-kube-api-access-bnpk6\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.733049 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-run-httpd\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.733072 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.733937 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-log-httpd\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.733976 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-run-httpd\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.739810 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-scripts\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.740107 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.743782 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-config-data\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.748480 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.756299 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnpk6\" (UniqueName: \"kubernetes.io/projected/dfadac38-b67f-4652-a05e-1f9ee873ab5b-kube-api-access-bnpk6\") pod \"ceilometer-0\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " pod="openstack/ceilometer-0" Oct 10 16:29:42 crc kubenswrapper[4788]: I1010 16:29:42.839539 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:29:43 crc kubenswrapper[4788]: I1010 16:29:43.352778 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:29:43 crc kubenswrapper[4788]: W1010 16:29:43.377880 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfadac38_b67f_4652_a05e_1f9ee873ab5b.slice/crio-c0f32dacdf9c5d1490edcb392aafbb14f093ce704ef2d639d3ad52db1d6d509d WatchSource:0}: Error finding container c0f32dacdf9c5d1490edcb392aafbb14f093ce704ef2d639d3ad52db1d6d509d: Status 404 returned error can't find the container with id c0f32dacdf9c5d1490edcb392aafbb14f093ce704ef2d639d3ad52db1d6d509d Oct 10 16:29:44 crc kubenswrapper[4788]: I1010 16:29:44.316452 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerStarted","Data":"0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1"} Oct 10 16:29:44 crc kubenswrapper[4788]: I1010 16:29:44.316925 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerStarted","Data":"c0f32dacdf9c5d1490edcb392aafbb14f093ce704ef2d639d3ad52db1d6d509d"} Oct 10 16:29:45 crc kubenswrapper[4788]: I1010 16:29:45.341435 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerStarted","Data":"0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575"} Oct 10 16:29:46 crc kubenswrapper[4788]: I1010 16:29:46.354665 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerStarted","Data":"857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b"} Oct 10 16:29:48 crc kubenswrapper[4788]: I1010 16:29:48.381344 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerStarted","Data":"ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836"} Oct 10 16:29:48 crc kubenswrapper[4788]: I1010 16:29:48.382087 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 16:29:48 crc kubenswrapper[4788]: I1010 16:29:48.409385 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.881945996 podStartE2EDuration="6.409358446s" podCreationTimestamp="2025-10-10 16:29:42 +0000 UTC" firstStartedPulling="2025-10-10 16:29:43.386425834 +0000 UTC m=+6285.836141382" lastFinishedPulling="2025-10-10 16:29:47.913838244 +0000 UTC m=+6290.363553832" observedRunningTime="2025-10-10 16:29:48.403835571 +0000 UTC m=+6290.853551129" watchObservedRunningTime="2025-10-10 16:29:48.409358446 +0000 UTC m=+6290.859074004" Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.033390 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-6j2kj"] Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.060449 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-6j2kj"] Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.667093 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-sgtqs"] Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.668600 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-sgtqs" Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.675150 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-sgtqs"] Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.805181 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m5b4\" (UniqueName: \"kubernetes.io/projected/eb4d0587-0164-4270-bb8d-93ec27bd3ab4-kube-api-access-6m5b4\") pod \"aodh-db-create-sgtqs\" (UID: \"eb4d0587-0164-4270-bb8d-93ec27bd3ab4\") " pod="openstack/aodh-db-create-sgtqs" Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.911126 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m5b4\" (UniqueName: \"kubernetes.io/projected/eb4d0587-0164-4270-bb8d-93ec27bd3ab4-kube-api-access-6m5b4\") pod \"aodh-db-create-sgtqs\" (UID: \"eb4d0587-0164-4270-bb8d-93ec27bd3ab4\") " pod="openstack/aodh-db-create-sgtqs" Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.948872 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m5b4\" (UniqueName: \"kubernetes.io/projected/eb4d0587-0164-4270-bb8d-93ec27bd3ab4-kube-api-access-6m5b4\") pod \"aodh-db-create-sgtqs\" (UID: \"eb4d0587-0164-4270-bb8d-93ec27bd3ab4\") " pod="openstack/aodh-db-create-sgtqs" Oct 10 16:29:53 crc kubenswrapper[4788]: I1010 16:29:53.988785 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-sgtqs" Oct 10 16:29:54 crc kubenswrapper[4788]: I1010 16:29:54.258244 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3021cf2d-3cf9-42c8-8a45-52bf67618e27" path="/var/lib/kubelet/pods/3021cf2d-3cf9-42c8-8a45-52bf67618e27/volumes" Oct 10 16:29:54 crc kubenswrapper[4788]: I1010 16:29:54.552645 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-sgtqs"] Oct 10 16:29:54 crc kubenswrapper[4788]: W1010 16:29:54.563511 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb4d0587_0164_4270_bb8d_93ec27bd3ab4.slice/crio-0761f12e8b15d45574750a20f0ede89630ce25ec27bd70cf947e1f4308396190 WatchSource:0}: Error finding container 0761f12e8b15d45574750a20f0ede89630ce25ec27bd70cf947e1f4308396190: Status 404 returned error can't find the container with id 0761f12e8b15d45574750a20f0ede89630ce25ec27bd70cf947e1f4308396190 Oct 10 16:29:55 crc kubenswrapper[4788]: I1010 16:29:55.449626 4788 generic.go:334] "Generic (PLEG): container finished" podID="eb4d0587-0164-4270-bb8d-93ec27bd3ab4" containerID="c5230e76d30353948a941e616ed001f193aa6353dfe70040b02810c412374d12" exitCode=0 Oct 10 16:29:55 crc kubenswrapper[4788]: I1010 16:29:55.449708 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-sgtqs" event={"ID":"eb4d0587-0164-4270-bb8d-93ec27bd3ab4","Type":"ContainerDied","Data":"c5230e76d30353948a941e616ed001f193aa6353dfe70040b02810c412374d12"} Oct 10 16:29:55 crc kubenswrapper[4788]: I1010 16:29:55.450015 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-sgtqs" event={"ID":"eb4d0587-0164-4270-bb8d-93ec27bd3ab4","Type":"ContainerStarted","Data":"0761f12e8b15d45574750a20f0ede89630ce25ec27bd70cf947e1f4308396190"} Oct 10 16:29:56 crc kubenswrapper[4788]: I1010 16:29:56.864697 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-sgtqs" Oct 10 16:29:56 crc kubenswrapper[4788]: I1010 16:29:56.986875 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m5b4\" (UniqueName: \"kubernetes.io/projected/eb4d0587-0164-4270-bb8d-93ec27bd3ab4-kube-api-access-6m5b4\") pod \"eb4d0587-0164-4270-bb8d-93ec27bd3ab4\" (UID: \"eb4d0587-0164-4270-bb8d-93ec27bd3ab4\") " Oct 10 16:29:56 crc kubenswrapper[4788]: I1010 16:29:56.994855 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb4d0587-0164-4270-bb8d-93ec27bd3ab4-kube-api-access-6m5b4" (OuterVolumeSpecName: "kube-api-access-6m5b4") pod "eb4d0587-0164-4270-bb8d-93ec27bd3ab4" (UID: "eb4d0587-0164-4270-bb8d-93ec27bd3ab4"). InnerVolumeSpecName "kube-api-access-6m5b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:29:57 crc kubenswrapper[4788]: I1010 16:29:57.091575 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m5b4\" (UniqueName: \"kubernetes.io/projected/eb4d0587-0164-4270-bb8d-93ec27bd3ab4-kube-api-access-6m5b4\") on node \"crc\" DevicePath \"\"" Oct 10 16:29:57 crc kubenswrapper[4788]: I1010 16:29:57.472296 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-sgtqs" event={"ID":"eb4d0587-0164-4270-bb8d-93ec27bd3ab4","Type":"ContainerDied","Data":"0761f12e8b15d45574750a20f0ede89630ce25ec27bd70cf947e1f4308396190"} Oct 10 16:29:57 crc kubenswrapper[4788]: I1010 16:29:57.472329 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-sgtqs" Oct 10 16:29:57 crc kubenswrapper[4788]: I1010 16:29:57.472343 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0761f12e8b15d45574750a20f0ede89630ce25ec27bd70cf947e1f4308396190" Oct 10 16:29:59 crc kubenswrapper[4788]: I1010 16:29:59.406410 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:29:59 crc kubenswrapper[4788]: I1010 16:29:59.407619 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:29:59 crc kubenswrapper[4788]: I1010 16:29:59.407721 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:29:59 crc kubenswrapper[4788]: I1010 16:29:59.409456 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:29:59 crc kubenswrapper[4788]: I1010 16:29:59.409579 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" gracePeriod=600 Oct 10 16:29:59 crc kubenswrapper[4788]: E1010 16:29:59.533981 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.158608 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72"] Oct 10 16:30:00 crc kubenswrapper[4788]: E1010 16:30:00.159851 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb4d0587-0164-4270-bb8d-93ec27bd3ab4" containerName="mariadb-database-create" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.159878 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb4d0587-0164-4270-bb8d-93ec27bd3ab4" containerName="mariadb-database-create" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.160133 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb4d0587-0164-4270-bb8d-93ec27bd3ab4" containerName="mariadb-database-create" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.161602 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.164265 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.164290 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.177611 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72"] Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.259957 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b448e2-395e-4e6b-9aa6-f029920da644-secret-volume\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.260363 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b448e2-395e-4e6b-9aa6-f029920da644-config-volume\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.260414 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkbt6\" (UniqueName: \"kubernetes.io/projected/59b448e2-395e-4e6b-9aa6-f029920da644-kube-api-access-bkbt6\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.364836 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b448e2-395e-4e6b-9aa6-f029920da644-config-volume\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.364895 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkbt6\" (UniqueName: \"kubernetes.io/projected/59b448e2-395e-4e6b-9aa6-f029920da644-kube-api-access-bkbt6\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.365125 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b448e2-395e-4e6b-9aa6-f029920da644-secret-volume\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.366206 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b448e2-395e-4e6b-9aa6-f029920da644-config-volume\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.372189 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b448e2-395e-4e6b-9aa6-f029920da644-secret-volume\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.386423 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkbt6\" (UniqueName: \"kubernetes.io/projected/59b448e2-395e-4e6b-9aa6-f029920da644-kube-api-access-bkbt6\") pod \"collect-profiles-29335230-wcr72\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.489284 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.508772 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" exitCode=0 Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.508874 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3"} Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.508943 4788 scope.go:117] "RemoveContainer" containerID="e510a005acb06733b8f662e2f362c13e2ba92b9f626ca5375903dde617f1f2fe" Oct 10 16:30:00 crc kubenswrapper[4788]: I1010 16:30:00.510661 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:30:00 crc kubenswrapper[4788]: E1010 16:30:00.511381 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:30:01 crc kubenswrapper[4788]: I1010 16:30:01.023766 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72"] Oct 10 16:30:01 crc kubenswrapper[4788]: W1010 16:30:01.032642 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59b448e2_395e_4e6b_9aa6_f029920da644.slice/crio-12ff3c6731f46af8dff3f457f0208955458cac86e69dc08bff4c7f78afa34331 WatchSource:0}: Error finding container 12ff3c6731f46af8dff3f457f0208955458cac86e69dc08bff4c7f78afa34331: Status 404 returned error can't find the container with id 12ff3c6731f46af8dff3f457f0208955458cac86e69dc08bff4c7f78afa34331 Oct 10 16:30:01 crc kubenswrapper[4788]: I1010 16:30:01.523349 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" event={"ID":"59b448e2-395e-4e6b-9aa6-f029920da644","Type":"ContainerStarted","Data":"16ef81a728b43af8397b8c84c8be6a4fb55918f928dd006e91bb1705dbed929f"} Oct 10 16:30:01 crc kubenswrapper[4788]: I1010 16:30:01.523822 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" event={"ID":"59b448e2-395e-4e6b-9aa6-f029920da644","Type":"ContainerStarted","Data":"12ff3c6731f46af8dff3f457f0208955458cac86e69dc08bff4c7f78afa34331"} Oct 10 16:30:01 crc kubenswrapper[4788]: I1010 16:30:01.553797 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" podStartSLOduration=1.553762547 podStartE2EDuration="1.553762547s" podCreationTimestamp="2025-10-10 16:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:30:01.542079629 +0000 UTC m=+6303.991795207" watchObservedRunningTime="2025-10-10 16:30:01.553762547 +0000 UTC m=+6304.003478105" Oct 10 16:30:02 crc kubenswrapper[4788]: I1010 16:30:02.547704 4788 generic.go:334] "Generic (PLEG): container finished" podID="59b448e2-395e-4e6b-9aa6-f029920da644" containerID="16ef81a728b43af8397b8c84c8be6a4fb55918f928dd006e91bb1705dbed929f" exitCode=0 Oct 10 16:30:02 crc kubenswrapper[4788]: I1010 16:30:02.548802 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" event={"ID":"59b448e2-395e-4e6b-9aa6-f029920da644","Type":"ContainerDied","Data":"16ef81a728b43af8397b8c84c8be6a4fb55918f928dd006e91bb1705dbed929f"} Oct 10 16:30:03 crc kubenswrapper[4788]: I1010 16:30:03.768639 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-d613-account-create-99s9t"] Oct 10 16:30:03 crc kubenswrapper[4788]: I1010 16:30:03.770029 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d613-account-create-99s9t" Oct 10 16:30:03 crc kubenswrapper[4788]: I1010 16:30:03.773314 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 10 16:30:03 crc kubenswrapper[4788]: I1010 16:30:03.792328 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-d613-account-create-99s9t"] Oct 10 16:30:03 crc kubenswrapper[4788]: I1010 16:30:03.873992 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghk5h\" (UniqueName: \"kubernetes.io/projected/2c8a2701-d615-41c9-bf5a-ead64d030646-kube-api-access-ghk5h\") pod \"aodh-d613-account-create-99s9t\" (UID: \"2c8a2701-d615-41c9-bf5a-ead64d030646\") " pod="openstack/aodh-d613-account-create-99s9t" Oct 10 16:30:03 crc kubenswrapper[4788]: I1010 16:30:03.975786 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghk5h\" (UniqueName: \"kubernetes.io/projected/2c8a2701-d615-41c9-bf5a-ead64d030646-kube-api-access-ghk5h\") pod \"aodh-d613-account-create-99s9t\" (UID: \"2c8a2701-d615-41c9-bf5a-ead64d030646\") " pod="openstack/aodh-d613-account-create-99s9t" Oct 10 16:30:03 crc kubenswrapper[4788]: I1010 16:30:03.994766 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghk5h\" (UniqueName: \"kubernetes.io/projected/2c8a2701-d615-41c9-bf5a-ead64d030646-kube-api-access-ghk5h\") pod \"aodh-d613-account-create-99s9t\" (UID: \"2c8a2701-d615-41c9-bf5a-ead64d030646\") " pod="openstack/aodh-d613-account-create-99s9t" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.040550 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.108778 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d613-account-create-99s9t" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.183668 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b448e2-395e-4e6b-9aa6-f029920da644-secret-volume\") pod \"59b448e2-395e-4e6b-9aa6-f029920da644\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.184163 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b448e2-395e-4e6b-9aa6-f029920da644-config-volume\") pod \"59b448e2-395e-4e6b-9aa6-f029920da644\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.184256 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkbt6\" (UniqueName: \"kubernetes.io/projected/59b448e2-395e-4e6b-9aa6-f029920da644-kube-api-access-bkbt6\") pod \"59b448e2-395e-4e6b-9aa6-f029920da644\" (UID: \"59b448e2-395e-4e6b-9aa6-f029920da644\") " Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.185003 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59b448e2-395e-4e6b-9aa6-f029920da644-config-volume" (OuterVolumeSpecName: "config-volume") pod "59b448e2-395e-4e6b-9aa6-f029920da644" (UID: "59b448e2-395e-4e6b-9aa6-f029920da644"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.190414 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b448e2-395e-4e6b-9aa6-f029920da644-kube-api-access-bkbt6" (OuterVolumeSpecName: "kube-api-access-bkbt6") pod "59b448e2-395e-4e6b-9aa6-f029920da644" (UID: "59b448e2-395e-4e6b-9aa6-f029920da644"). InnerVolumeSpecName "kube-api-access-bkbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.198984 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59b448e2-395e-4e6b-9aa6-f029920da644-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "59b448e2-395e-4e6b-9aa6-f029920da644" (UID: "59b448e2-395e-4e6b-9aa6-f029920da644"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.286625 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b448e2-395e-4e6b-9aa6-f029920da644-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.286665 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkbt6\" (UniqueName: \"kubernetes.io/projected/59b448e2-395e-4e6b-9aa6-f029920da644-kube-api-access-bkbt6\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.286675 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b448e2-395e-4e6b-9aa6-f029920da644-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.543792 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-d613-account-create-99s9t"] Oct 10 16:30:04 crc kubenswrapper[4788]: W1010 16:30:04.545097 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c8a2701_d615_41c9_bf5a_ead64d030646.slice/crio-f57fb8cbc6d94549364e6ab524b122ea7986f9bdd7e07708bebcaa2b334cd287 WatchSource:0}: Error finding container f57fb8cbc6d94549364e6ab524b122ea7986f9bdd7e07708bebcaa2b334cd287: Status 404 returned error can't find the container with id f57fb8cbc6d94549364e6ab524b122ea7986f9bdd7e07708bebcaa2b334cd287 Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.572902 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d613-account-create-99s9t" event={"ID":"2c8a2701-d615-41c9-bf5a-ead64d030646","Type":"ContainerStarted","Data":"f57fb8cbc6d94549364e6ab524b122ea7986f9bdd7e07708bebcaa2b334cd287"} Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.575164 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" event={"ID":"59b448e2-395e-4e6b-9aa6-f029920da644","Type":"ContainerDied","Data":"12ff3c6731f46af8dff3f457f0208955458cac86e69dc08bff4c7f78afa34331"} Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.575226 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12ff3c6731f46af8dff3f457f0208955458cac86e69dc08bff4c7f78afa34331" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.575287 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72" Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.604195 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd"] Oct 10 16:30:04 crc kubenswrapper[4788]: I1010 16:30:04.612490 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335185-7r4vd"] Oct 10 16:30:05 crc kubenswrapper[4788]: I1010 16:30:05.587042 4788 generic.go:334] "Generic (PLEG): container finished" podID="2c8a2701-d615-41c9-bf5a-ead64d030646" containerID="2cc1764388ca1ef85001d35b8a632684ea98a4e26a0a852d7a4042529fdd300c" exitCode=0 Oct 10 16:30:05 crc kubenswrapper[4788]: I1010 16:30:05.587091 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d613-account-create-99s9t" event={"ID":"2c8a2701-d615-41c9-bf5a-ead64d030646","Type":"ContainerDied","Data":"2cc1764388ca1ef85001d35b8a632684ea98a4e26a0a852d7a4042529fdd300c"} Oct 10 16:30:06 crc kubenswrapper[4788]: I1010 16:30:06.248180 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ee73fd4-d24a-498b-b48a-aca42501618f" path="/var/lib/kubelet/pods/5ee73fd4-d24a-498b-b48a-aca42501618f/volumes" Oct 10 16:30:07 crc kubenswrapper[4788]: I1010 16:30:07.053383 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d613-account-create-99s9t" Oct 10 16:30:07 crc kubenswrapper[4788]: I1010 16:30:07.143695 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghk5h\" (UniqueName: \"kubernetes.io/projected/2c8a2701-d615-41c9-bf5a-ead64d030646-kube-api-access-ghk5h\") pod \"2c8a2701-d615-41c9-bf5a-ead64d030646\" (UID: \"2c8a2701-d615-41c9-bf5a-ead64d030646\") " Oct 10 16:30:07 crc kubenswrapper[4788]: I1010 16:30:07.149539 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c8a2701-d615-41c9-bf5a-ead64d030646-kube-api-access-ghk5h" (OuterVolumeSpecName: "kube-api-access-ghk5h") pod "2c8a2701-d615-41c9-bf5a-ead64d030646" (UID: "2c8a2701-d615-41c9-bf5a-ead64d030646"). InnerVolumeSpecName "kube-api-access-ghk5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:30:07 crc kubenswrapper[4788]: I1010 16:30:07.245646 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghk5h\" (UniqueName: \"kubernetes.io/projected/2c8a2701-d615-41c9-bf5a-ead64d030646-kube-api-access-ghk5h\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:07 crc kubenswrapper[4788]: I1010 16:30:07.619007 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d613-account-create-99s9t" event={"ID":"2c8a2701-d615-41c9-bf5a-ead64d030646","Type":"ContainerDied","Data":"f57fb8cbc6d94549364e6ab524b122ea7986f9bdd7e07708bebcaa2b334cd287"} Oct 10 16:30:07 crc kubenswrapper[4788]: I1010 16:30:07.619050 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f57fb8cbc6d94549364e6ab524b122ea7986f9bdd7e07708bebcaa2b334cd287" Oct 10 16:30:07 crc kubenswrapper[4788]: I1010 16:30:07.619088 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d613-account-create-99s9t" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.204739 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-w4rcs"] Oct 10 16:30:09 crc kubenswrapper[4788]: E1010 16:30:09.205546 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b448e2-395e-4e6b-9aa6-f029920da644" containerName="collect-profiles" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.205567 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b448e2-395e-4e6b-9aa6-f029920da644" containerName="collect-profiles" Oct 10 16:30:09 crc kubenswrapper[4788]: E1010 16:30:09.205610 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c8a2701-d615-41c9-bf5a-ead64d030646" containerName="mariadb-account-create" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.205621 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c8a2701-d615-41c9-bf5a-ead64d030646" containerName="mariadb-account-create" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.205909 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c8a2701-d615-41c9-bf5a-ead64d030646" containerName="mariadb-account-create" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.205934 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b448e2-395e-4e6b-9aa6-f029920da644" containerName="collect-profiles" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.206799 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.215014 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-w4rcs"] Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.215357 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.215443 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-ql2qh" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.215794 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.288751 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-combined-ca-bundle\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.288932 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v26zx\" (UniqueName: \"kubernetes.io/projected/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-kube-api-access-v26zx\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.289075 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-config-data\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.289443 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-scripts\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.392319 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-scripts\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.392584 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-combined-ca-bundle\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.392653 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v26zx\" (UniqueName: \"kubernetes.io/projected/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-kube-api-access-v26zx\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.392679 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-config-data\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.398035 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-scripts\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.408281 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-config-data\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.417644 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-combined-ca-bundle\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.418844 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v26zx\" (UniqueName: \"kubernetes.io/projected/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-kube-api-access-v26zx\") pod \"aodh-db-sync-w4rcs\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:09 crc kubenswrapper[4788]: I1010 16:30:09.545246 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:10 crc kubenswrapper[4788]: I1010 16:30:10.091942 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-w4rcs"] Oct 10 16:30:10 crc kubenswrapper[4788]: I1010 16:30:10.648248 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-w4rcs" event={"ID":"d180f85a-fb1a-44ef-83f0-dae3057fb6e6","Type":"ContainerStarted","Data":"14884ae19e42474547b1fa35e9d0627f76e9841c43a5081946f710251e1f2b86"} Oct 10 16:30:12 crc kubenswrapper[4788]: I1010 16:30:12.850600 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 16:30:15 crc kubenswrapper[4788]: I1010 16:30:15.235091 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:30:15 crc kubenswrapper[4788]: E1010 16:30:15.236078 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:30:15 crc kubenswrapper[4788]: I1010 16:30:15.654309 4788 scope.go:117] "RemoveContainer" containerID="6634e9f5cb7db0dfb8b5eeb1de0480036835200f9e1d838d2abf7b09b7bed166" Oct 10 16:30:17 crc kubenswrapper[4788]: I1010 16:30:17.411529 4788 scope.go:117] "RemoveContainer" containerID="7e3baf9d30c8a0b8a0496f374a184959c6cee6152f43000611a26b5dcd775b30" Oct 10 16:30:17 crc kubenswrapper[4788]: I1010 16:30:17.469377 4788 scope.go:117] "RemoveContainer" containerID="9a75079f37187cb3ddd27b0c52c050d7a897a0542fd238a327dfd4517ee27cfa" Oct 10 16:30:17 crc kubenswrapper[4788]: I1010 16:30:17.610106 4788 scope.go:117] "RemoveContainer" containerID="4d1039c1755f06c7ab69ba240ecbd4b8f78e464881ef87611b32129e4152d420" Oct 10 16:30:17 crc kubenswrapper[4788]: I1010 16:30:17.693120 4788 scope.go:117] "RemoveContainer" containerID="b34b745cdb3e93eff8625aece80c287229820a4544266b653fcdd0b1cf04fd44" Oct 10 16:30:17 crc kubenswrapper[4788]: I1010 16:30:17.718450 4788 scope.go:117] "RemoveContainer" containerID="aab30d0014507e0985e587fd8c408952d9eaf41ca860494081804986a1d0230e" Oct 10 16:30:18 crc kubenswrapper[4788]: I1010 16:30:18.743702 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-w4rcs" event={"ID":"d180f85a-fb1a-44ef-83f0-dae3057fb6e6","Type":"ContainerStarted","Data":"e8d98220b03bc653d3b269606c923c9f28f099265fd0051e87d118cbf6a3932c"} Oct 10 16:30:18 crc kubenswrapper[4788]: I1010 16:30:18.763812 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-w4rcs" podStartSLOduration=2.489472574 podStartE2EDuration="9.763793367s" podCreationTimestamp="2025-10-10 16:30:09 +0000 UTC" firstStartedPulling="2025-10-10 16:30:10.1958111 +0000 UTC m=+6312.645526648" lastFinishedPulling="2025-10-10 16:30:17.470131893 +0000 UTC m=+6319.919847441" observedRunningTime="2025-10-10 16:30:18.756362821 +0000 UTC m=+6321.206078369" watchObservedRunningTime="2025-10-10 16:30:18.763793367 +0000 UTC m=+6321.213508915" Oct 10 16:30:20 crc kubenswrapper[4788]: I1010 16:30:20.765557 4788 generic.go:334] "Generic (PLEG): container finished" podID="d180f85a-fb1a-44ef-83f0-dae3057fb6e6" containerID="e8d98220b03bc653d3b269606c923c9f28f099265fd0051e87d118cbf6a3932c" exitCode=0 Oct 10 16:30:20 crc kubenswrapper[4788]: I1010 16:30:20.765599 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-w4rcs" event={"ID":"d180f85a-fb1a-44ef-83f0-dae3057fb6e6","Type":"ContainerDied","Data":"e8d98220b03bc653d3b269606c923c9f28f099265fd0051e87d118cbf6a3932c"} Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.224812 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.303959 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-combined-ca-bundle\") pod \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.304108 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-scripts\") pod \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.304292 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v26zx\" (UniqueName: \"kubernetes.io/projected/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-kube-api-access-v26zx\") pod \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.304412 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-config-data\") pod \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\" (UID: \"d180f85a-fb1a-44ef-83f0-dae3057fb6e6\") " Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.310632 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-scripts" (OuterVolumeSpecName: "scripts") pod "d180f85a-fb1a-44ef-83f0-dae3057fb6e6" (UID: "d180f85a-fb1a-44ef-83f0-dae3057fb6e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.310743 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-kube-api-access-v26zx" (OuterVolumeSpecName: "kube-api-access-v26zx") pod "d180f85a-fb1a-44ef-83f0-dae3057fb6e6" (UID: "d180f85a-fb1a-44ef-83f0-dae3057fb6e6"). InnerVolumeSpecName "kube-api-access-v26zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.334956 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-config-data" (OuterVolumeSpecName: "config-data") pod "d180f85a-fb1a-44ef-83f0-dae3057fb6e6" (UID: "d180f85a-fb1a-44ef-83f0-dae3057fb6e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.335274 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d180f85a-fb1a-44ef-83f0-dae3057fb6e6" (UID: "d180f85a-fb1a-44ef-83f0-dae3057fb6e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.406657 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v26zx\" (UniqueName: \"kubernetes.io/projected/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-kube-api-access-v26zx\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.406698 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.406707 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.406715 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d180f85a-fb1a-44ef-83f0-dae3057fb6e6-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.787023 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-w4rcs" event={"ID":"d180f85a-fb1a-44ef-83f0-dae3057fb6e6","Type":"ContainerDied","Data":"14884ae19e42474547b1fa35e9d0627f76e9841c43a5081946f710251e1f2b86"} Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.787381 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14884ae19e42474547b1fa35e9d0627f76e9841c43a5081946f710251e1f2b86" Oct 10 16:30:22 crc kubenswrapper[4788]: I1010 16:30:22.787110 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-w4rcs" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.781768 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 10 16:30:23 crc kubenswrapper[4788]: E1010 16:30:23.783127 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d180f85a-fb1a-44ef-83f0-dae3057fb6e6" containerName="aodh-db-sync" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.785310 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d180f85a-fb1a-44ef-83f0-dae3057fb6e6" containerName="aodh-db-sync" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.785604 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d180f85a-fb1a-44ef-83f0-dae3057fb6e6" containerName="aodh-db-sync" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.789470 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.792647 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.792827 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-ql2qh" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.793051 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.807005 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.842858 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkmvt\" (UniqueName: \"kubernetes.io/projected/7b0a366f-f03c-4607-8e50-6df8bf30d683-kube-api-access-vkmvt\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.842942 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-scripts\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.843122 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.843250 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-config-data\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.945101 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-config-data\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.945178 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkmvt\" (UniqueName: \"kubernetes.io/projected/7b0a366f-f03c-4607-8e50-6df8bf30d683-kube-api-access-vkmvt\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.945205 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-scripts\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.945311 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.951923 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-scripts\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.957019 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.957265 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b0a366f-f03c-4607-8e50-6df8bf30d683-config-data\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:23 crc kubenswrapper[4788]: I1010 16:30:23.985586 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkmvt\" (UniqueName: \"kubernetes.io/projected/7b0a366f-f03c-4607-8e50-6df8bf30d683-kube-api-access-vkmvt\") pod \"aodh-0\" (UID: \"7b0a366f-f03c-4607-8e50-6df8bf30d683\") " pod="openstack/aodh-0" Oct 10 16:30:24 crc kubenswrapper[4788]: I1010 16:30:24.114661 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 10 16:30:24 crc kubenswrapper[4788]: I1010 16:30:24.635091 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 10 16:30:24 crc kubenswrapper[4788]: I1010 16:30:24.805954 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7b0a366f-f03c-4607-8e50-6df8bf30d683","Type":"ContainerStarted","Data":"3710ad70a4dcb0dbd52a9b24ef1433f3fac77f5224374c47ebdb788cbb903fb3"} Oct 10 16:30:25 crc kubenswrapper[4788]: I1010 16:30:25.815763 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7b0a366f-f03c-4607-8e50-6df8bf30d683","Type":"ContainerStarted","Data":"fdd2edee484d671fcecb1c551a384d7124b6178fd5bfe0fddce6ba88ff9d0b3e"} Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.425053 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.425411 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="ceilometer-central-agent" containerID="cri-o://0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1" gracePeriod=30 Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.425438 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="proxy-httpd" containerID="cri-o://ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836" gracePeriod=30 Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.425577 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="ceilometer-notification-agent" containerID="cri-o://0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575" gracePeriod=30 Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.425903 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="sg-core" containerID="cri-o://857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b" gracePeriod=30 Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.827274 4788 generic.go:334] "Generic (PLEG): container finished" podID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerID="ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836" exitCode=0 Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.827626 4788 generic.go:334] "Generic (PLEG): container finished" podID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerID="857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b" exitCode=2 Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.827314 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerDied","Data":"ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836"} Oct 10 16:30:26 crc kubenswrapper[4788]: I1010 16:30:26.827660 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerDied","Data":"857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b"} Oct 10 16:30:27 crc kubenswrapper[4788]: I1010 16:30:27.233593 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:30:27 crc kubenswrapper[4788]: E1010 16:30:27.234080 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:30:27 crc kubenswrapper[4788]: I1010 16:30:27.844291 4788 generic.go:334] "Generic (PLEG): container finished" podID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerID="0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1" exitCode=0 Oct 10 16:30:27 crc kubenswrapper[4788]: I1010 16:30:27.844362 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerDied","Data":"0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1"} Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.476090 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.564088 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-sg-core-conf-yaml\") pod \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.564192 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-scripts\") pod \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.564255 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnpk6\" (UniqueName: \"kubernetes.io/projected/dfadac38-b67f-4652-a05e-1f9ee873ab5b-kube-api-access-bnpk6\") pod \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.564323 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-combined-ca-bundle\") pod \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.564379 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-run-httpd\") pod \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.564432 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-config-data\") pod \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.564453 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-log-httpd\") pod \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\" (UID: \"dfadac38-b67f-4652-a05e-1f9ee873ab5b\") " Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.565074 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dfadac38-b67f-4652-a05e-1f9ee873ab5b" (UID: "dfadac38-b67f-4652-a05e-1f9ee873ab5b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.565282 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dfadac38-b67f-4652-a05e-1f9ee873ab5b" (UID: "dfadac38-b67f-4652-a05e-1f9ee873ab5b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.570162 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-scripts" (OuterVolumeSpecName: "scripts") pod "dfadac38-b67f-4652-a05e-1f9ee873ab5b" (UID: "dfadac38-b67f-4652-a05e-1f9ee873ab5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.570501 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfadac38-b67f-4652-a05e-1f9ee873ab5b-kube-api-access-bnpk6" (OuterVolumeSpecName: "kube-api-access-bnpk6") pod "dfadac38-b67f-4652-a05e-1f9ee873ab5b" (UID: "dfadac38-b67f-4652-a05e-1f9ee873ab5b"). InnerVolumeSpecName "kube-api-access-bnpk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.596070 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dfadac38-b67f-4652-a05e-1f9ee873ab5b" (UID: "dfadac38-b67f-4652-a05e-1f9ee873ab5b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.663659 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfadac38-b67f-4652-a05e-1f9ee873ab5b" (UID: "dfadac38-b67f-4652-a05e-1f9ee873ab5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.666930 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.666959 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.666969 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.666979 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnpk6\" (UniqueName: \"kubernetes.io/projected/dfadac38-b67f-4652-a05e-1f9ee873ab5b-kube-api-access-bnpk6\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.666990 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.666998 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dfadac38-b67f-4652-a05e-1f9ee873ab5b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.684492 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-config-data" (OuterVolumeSpecName: "config-data") pod "dfadac38-b67f-4652-a05e-1f9ee873ab5b" (UID: "dfadac38-b67f-4652-a05e-1f9ee873ab5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.768850 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfadac38-b67f-4652-a05e-1f9ee873ab5b-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.860087 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7b0a366f-f03c-4607-8e50-6df8bf30d683","Type":"ContainerStarted","Data":"b9698022c99d64d21433a063e9be1fb2bdb425e3439536a8dd35136be39fcc89"} Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.864301 4788 generic.go:334] "Generic (PLEG): container finished" podID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerID="0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575" exitCode=0 Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.864358 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerDied","Data":"0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575"} Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.864378 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.864405 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dfadac38-b67f-4652-a05e-1f9ee873ab5b","Type":"ContainerDied","Data":"c0f32dacdf9c5d1490edcb392aafbb14f093ce704ef2d639d3ad52db1d6d509d"} Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.864430 4788 scope.go:117] "RemoveContainer" containerID="ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.912120 4788 scope.go:117] "RemoveContainer" containerID="857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.925444 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.938222 4788 scope.go:117] "RemoveContainer" containerID="0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.940458 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.954995 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:30:28 crc kubenswrapper[4788]: E1010 16:30:28.955414 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="proxy-httpd" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.955431 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="proxy-httpd" Oct 10 16:30:28 crc kubenswrapper[4788]: E1010 16:30:28.955449 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="ceilometer-notification-agent" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.955456 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="ceilometer-notification-agent" Oct 10 16:30:28 crc kubenswrapper[4788]: E1010 16:30:28.955483 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="sg-core" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.955489 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="sg-core" Oct 10 16:30:28 crc kubenswrapper[4788]: E1010 16:30:28.955512 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="ceilometer-central-agent" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.955518 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="ceilometer-central-agent" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.955703 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="ceilometer-notification-agent" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.955721 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="ceilometer-central-agent" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.955735 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="proxy-httpd" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.955758 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" containerName="sg-core" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.957568 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.959520 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.959814 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.964726 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:30:28 crc kubenswrapper[4788]: I1010 16:30:28.990050 4788 scope.go:117] "RemoveContainer" containerID="0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.013087 4788 scope.go:117] "RemoveContainer" containerID="ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836" Oct 10 16:30:29 crc kubenswrapper[4788]: E1010 16:30:29.013528 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836\": container with ID starting with ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836 not found: ID does not exist" containerID="ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.013559 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836"} err="failed to get container status \"ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836\": rpc error: code = NotFound desc = could not find container \"ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836\": container with ID starting with ac13eb6bd01681a74f7d69301f0ec1169db96cfe89fc1185da6d4df159b1f836 not found: ID does not exist" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.013577 4788 scope.go:117] "RemoveContainer" containerID="857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b" Oct 10 16:30:29 crc kubenswrapper[4788]: E1010 16:30:29.013827 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b\": container with ID starting with 857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b not found: ID does not exist" containerID="857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.013857 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b"} err="failed to get container status \"857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b\": rpc error: code = NotFound desc = could not find container \"857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b\": container with ID starting with 857339f96f2bc032f11f2ee91f1580a2471bc17d778460a80e54ce7b264a6a1b not found: ID does not exist" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.013874 4788 scope.go:117] "RemoveContainer" containerID="0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575" Oct 10 16:30:29 crc kubenswrapper[4788]: E1010 16:30:29.014265 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575\": container with ID starting with 0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575 not found: ID does not exist" containerID="0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.014293 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575"} err="failed to get container status \"0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575\": rpc error: code = NotFound desc = could not find container \"0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575\": container with ID starting with 0e4e41e1d45a232c32ca8f1e77177d5c86b141f1994986a184f91462296ea575 not found: ID does not exist" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.014304 4788 scope.go:117] "RemoveContainer" containerID="0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1" Oct 10 16:30:29 crc kubenswrapper[4788]: E1010 16:30:29.014545 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1\": container with ID starting with 0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1 not found: ID does not exist" containerID="0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.014574 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1"} err="failed to get container status \"0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1\": rpc error: code = NotFound desc = could not find container \"0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1\": container with ID starting with 0d4c6724349d8d677964359643a8b29070834b2203f52e0634ea8b2fe9b005e1 not found: ID does not exist" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.074155 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-config-data\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.074248 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-log-httpd\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.074267 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.074294 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-scripts\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.074389 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-run-httpd\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.074424 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z22w9\" (UniqueName: \"kubernetes.io/projected/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-kube-api-access-z22w9\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.074456 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.176024 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-config-data\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.176212 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-log-httpd\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.176252 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.176288 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-scripts\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.176409 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-run-httpd\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.176483 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z22w9\" (UniqueName: \"kubernetes.io/projected/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-kube-api-access-z22w9\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.176537 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.177287 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-log-httpd\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.177312 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-run-httpd\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.181273 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.181378 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-scripts\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.182287 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-config-data\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.183126 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.194364 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z22w9\" (UniqueName: \"kubernetes.io/projected/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-kube-api-access-z22w9\") pod \"ceilometer-0\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.299571 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.840578 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:30:29 crc kubenswrapper[4788]: I1010 16:30:29.879517 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerStarted","Data":"96e5a308e2c8d07b186678227c5000c0cf29be920b713f929426fbb23e34b8e5"} Oct 10 16:30:30 crc kubenswrapper[4788]: I1010 16:30:30.263881 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfadac38-b67f-4652-a05e-1f9ee873ab5b" path="/var/lib/kubelet/pods/dfadac38-b67f-4652-a05e-1f9ee873ab5b/volumes" Oct 10 16:30:30 crc kubenswrapper[4788]: I1010 16:30:30.892238 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerStarted","Data":"7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c"} Oct 10 16:30:31 crc kubenswrapper[4788]: I1010 16:30:31.907858 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerStarted","Data":"4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf"} Oct 10 16:30:31 crc kubenswrapper[4788]: I1010 16:30:31.910453 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7b0a366f-f03c-4607-8e50-6df8bf30d683","Type":"ContainerStarted","Data":"67556ff91286f2d706ec5d6f09694ec24437ca1a3db1338e3bee5dfc2d21701b"} Oct 10 16:30:32 crc kubenswrapper[4788]: I1010 16:30:32.934359 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerStarted","Data":"ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329"} Oct 10 16:30:34 crc kubenswrapper[4788]: I1010 16:30:34.977624 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7b0a366f-f03c-4607-8e50-6df8bf30d683","Type":"ContainerStarted","Data":"ef252412f4643a83b5bde6f80ceda567923fe9a01f23b1e921097194c9f05ab1"} Oct 10 16:30:34 crc kubenswrapper[4788]: I1010 16:30:34.982868 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerStarted","Data":"0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b"} Oct 10 16:30:34 crc kubenswrapper[4788]: I1010 16:30:34.983070 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 16:30:35 crc kubenswrapper[4788]: I1010 16:30:35.029895 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.860044658 podStartE2EDuration="12.029872585s" podCreationTimestamp="2025-10-10 16:30:23 +0000 UTC" firstStartedPulling="2025-10-10 16:30:24.638807316 +0000 UTC m=+6327.088522854" lastFinishedPulling="2025-10-10 16:30:33.808635223 +0000 UTC m=+6336.258350781" observedRunningTime="2025-10-10 16:30:35.027912763 +0000 UTC m=+6337.477628311" watchObservedRunningTime="2025-10-10 16:30:35.029872585 +0000 UTC m=+6337.479588133" Oct 10 16:30:35 crc kubenswrapper[4788]: I1010 16:30:35.072867 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.960854787 podStartE2EDuration="7.072843089s" podCreationTimestamp="2025-10-10 16:30:28 +0000 UTC" firstStartedPulling="2025-10-10 16:30:29.855472673 +0000 UTC m=+6332.305188231" lastFinishedPulling="2025-10-10 16:30:33.967460985 +0000 UTC m=+6336.417176533" observedRunningTime="2025-10-10 16:30:35.063119743 +0000 UTC m=+6337.512835321" watchObservedRunningTime="2025-10-10 16:30:35.072843089 +0000 UTC m=+6337.522558647" Oct 10 16:30:35 crc kubenswrapper[4788]: I1010 16:30:35.099377 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-cdcgn"] Oct 10 16:30:35 crc kubenswrapper[4788]: I1010 16:30:35.109809 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-cdcgn"] Oct 10 16:30:36 crc kubenswrapper[4788]: I1010 16:30:36.256247 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9be368-d869-476f-a24e-682a3ec8821d" path="/var/lib/kubelet/pods/de9be368-d869-476f-a24e-682a3ec8821d/volumes" Oct 10 16:30:41 crc kubenswrapper[4788]: I1010 16:30:41.043085 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-mk5cn"] Oct 10 16:30:41 crc kubenswrapper[4788]: I1010 16:30:41.045612 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mk5cn" Oct 10 16:30:41 crc kubenswrapper[4788]: I1010 16:30:41.088600 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mk5cn"] Oct 10 16:30:41 crc kubenswrapper[4788]: I1010 16:30:41.191591 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/a38bd59d-0d0f-4060-8a9f-09934dfbe079-kube-api-access-hhgz5\") pod \"manila-db-create-mk5cn\" (UID: \"a38bd59d-0d0f-4060-8a9f-09934dfbe079\") " pod="openstack/manila-db-create-mk5cn" Oct 10 16:30:41 crc kubenswrapper[4788]: I1010 16:30:41.293992 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/a38bd59d-0d0f-4060-8a9f-09934dfbe079-kube-api-access-hhgz5\") pod \"manila-db-create-mk5cn\" (UID: \"a38bd59d-0d0f-4060-8a9f-09934dfbe079\") " pod="openstack/manila-db-create-mk5cn" Oct 10 16:30:41 crc kubenswrapper[4788]: I1010 16:30:41.314213 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/a38bd59d-0d0f-4060-8a9f-09934dfbe079-kube-api-access-hhgz5\") pod \"manila-db-create-mk5cn\" (UID: \"a38bd59d-0d0f-4060-8a9f-09934dfbe079\") " pod="openstack/manila-db-create-mk5cn" Oct 10 16:30:41 crc kubenswrapper[4788]: I1010 16:30:41.390533 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mk5cn" Oct 10 16:30:41 crc kubenswrapper[4788]: I1010 16:30:41.903944 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mk5cn"] Oct 10 16:30:41 crc kubenswrapper[4788]: W1010 16:30:41.908664 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda38bd59d_0d0f_4060_8a9f_09934dfbe079.slice/crio-656523295cf05bbfec417fb4aa9416bb61c45868a94457dcc3a9c828176cc4f0 WatchSource:0}: Error finding container 656523295cf05bbfec417fb4aa9416bb61c45868a94457dcc3a9c828176cc4f0: Status 404 returned error can't find the container with id 656523295cf05bbfec417fb4aa9416bb61c45868a94457dcc3a9c828176cc4f0 Oct 10 16:30:42 crc kubenswrapper[4788]: I1010 16:30:42.080072 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mk5cn" event={"ID":"a38bd59d-0d0f-4060-8a9f-09934dfbe079","Type":"ContainerStarted","Data":"656523295cf05bbfec417fb4aa9416bb61c45868a94457dcc3a9c828176cc4f0"} Oct 10 16:30:42 crc kubenswrapper[4788]: I1010 16:30:42.234610 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:30:42 crc kubenswrapper[4788]: E1010 16:30:42.235090 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:30:43 crc kubenswrapper[4788]: I1010 16:30:43.093447 4788 generic.go:334] "Generic (PLEG): container finished" podID="a38bd59d-0d0f-4060-8a9f-09934dfbe079" containerID="6f698fb6c36499b5ad6e8f25c4ee094f2e76c24a0f36b7fdf9a1fc9869fc2816" exitCode=0 Oct 10 16:30:43 crc kubenswrapper[4788]: I1010 16:30:43.093563 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mk5cn" event={"ID":"a38bd59d-0d0f-4060-8a9f-09934dfbe079","Type":"ContainerDied","Data":"6f698fb6c36499b5ad6e8f25c4ee094f2e76c24a0f36b7fdf9a1fc9869fc2816"} Oct 10 16:30:44 crc kubenswrapper[4788]: I1010 16:30:44.518712 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mk5cn" Oct 10 16:30:44 crc kubenswrapper[4788]: I1010 16:30:44.672077 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/a38bd59d-0d0f-4060-8a9f-09934dfbe079-kube-api-access-hhgz5\") pod \"a38bd59d-0d0f-4060-8a9f-09934dfbe079\" (UID: \"a38bd59d-0d0f-4060-8a9f-09934dfbe079\") " Oct 10 16:30:44 crc kubenswrapper[4788]: I1010 16:30:44.680300 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a38bd59d-0d0f-4060-8a9f-09934dfbe079-kube-api-access-hhgz5" (OuterVolumeSpecName: "kube-api-access-hhgz5") pod "a38bd59d-0d0f-4060-8a9f-09934dfbe079" (UID: "a38bd59d-0d0f-4060-8a9f-09934dfbe079"). InnerVolumeSpecName "kube-api-access-hhgz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:30:44 crc kubenswrapper[4788]: I1010 16:30:44.774959 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/a38bd59d-0d0f-4060-8a9f-09934dfbe079-kube-api-access-hhgz5\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:45 crc kubenswrapper[4788]: I1010 16:30:45.054348 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-092d-account-create-dtpnc"] Oct 10 16:30:45 crc kubenswrapper[4788]: I1010 16:30:45.063694 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-092d-account-create-dtpnc"] Oct 10 16:30:45 crc kubenswrapper[4788]: I1010 16:30:45.116515 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mk5cn" event={"ID":"a38bd59d-0d0f-4060-8a9f-09934dfbe079","Type":"ContainerDied","Data":"656523295cf05bbfec417fb4aa9416bb61c45868a94457dcc3a9c828176cc4f0"} Oct 10 16:30:45 crc kubenswrapper[4788]: I1010 16:30:45.116563 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="656523295cf05bbfec417fb4aa9416bb61c45868a94457dcc3a9c828176cc4f0" Oct 10 16:30:45 crc kubenswrapper[4788]: I1010 16:30:45.116601 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mk5cn" Oct 10 16:30:46 crc kubenswrapper[4788]: I1010 16:30:46.246085 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d3d68a-818c-48af-bc70-09c7d11a9177" path="/var/lib/kubelet/pods/35d3d68a-818c-48af-bc70-09c7d11a9177/volumes" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.293095 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-3e16-account-create-xzt2v"] Oct 10 16:30:51 crc kubenswrapper[4788]: E1010 16:30:51.294061 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38bd59d-0d0f-4060-8a9f-09934dfbe079" containerName="mariadb-database-create" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.294074 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38bd59d-0d0f-4060-8a9f-09934dfbe079" containerName="mariadb-database-create" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.294310 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="a38bd59d-0d0f-4060-8a9f-09934dfbe079" containerName="mariadb-database-create" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.296487 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-3e16-account-create-xzt2v" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.299790 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.304877 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-3e16-account-create-xzt2v"] Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.359745 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9drq\" (UniqueName: \"kubernetes.io/projected/081f5e53-b6ea-41d5-ad7a-e1c11454af15-kube-api-access-r9drq\") pod \"manila-3e16-account-create-xzt2v\" (UID: \"081f5e53-b6ea-41d5-ad7a-e1c11454af15\") " pod="openstack/manila-3e16-account-create-xzt2v" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.461508 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9drq\" (UniqueName: \"kubernetes.io/projected/081f5e53-b6ea-41d5-ad7a-e1c11454af15-kube-api-access-r9drq\") pod \"manila-3e16-account-create-xzt2v\" (UID: \"081f5e53-b6ea-41d5-ad7a-e1c11454af15\") " pod="openstack/manila-3e16-account-create-xzt2v" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.479936 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9drq\" (UniqueName: \"kubernetes.io/projected/081f5e53-b6ea-41d5-ad7a-e1c11454af15-kube-api-access-r9drq\") pod \"manila-3e16-account-create-xzt2v\" (UID: \"081f5e53-b6ea-41d5-ad7a-e1c11454af15\") " pod="openstack/manila-3e16-account-create-xzt2v" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.625079 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-3e16-account-create-xzt2v" Oct 10 16:30:51 crc kubenswrapper[4788]: I1010 16:30:51.904596 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-3e16-account-create-xzt2v"] Oct 10 16:30:51 crc kubenswrapper[4788]: W1010 16:30:51.911501 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081f5e53_b6ea_41d5_ad7a_e1c11454af15.slice/crio-33de9e3c5a6f9508d58d58efab53f7b648d6364744af98da87fb74ab588ee04b WatchSource:0}: Error finding container 33de9e3c5a6f9508d58d58efab53f7b648d6364744af98da87fb74ab588ee04b: Status 404 returned error can't find the container with id 33de9e3c5a6f9508d58d58efab53f7b648d6364744af98da87fb74ab588ee04b Oct 10 16:30:52 crc kubenswrapper[4788]: I1010 16:30:52.186027 4788 generic.go:334] "Generic (PLEG): container finished" podID="081f5e53-b6ea-41d5-ad7a-e1c11454af15" containerID="a1ca0c753209c8fa0d85457ef7e97700181df10af5766ce36f9810d1b8ba7b21" exitCode=0 Oct 10 16:30:52 crc kubenswrapper[4788]: I1010 16:30:52.186085 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-3e16-account-create-xzt2v" event={"ID":"081f5e53-b6ea-41d5-ad7a-e1c11454af15","Type":"ContainerDied","Data":"a1ca0c753209c8fa0d85457ef7e97700181df10af5766ce36f9810d1b8ba7b21"} Oct 10 16:30:52 crc kubenswrapper[4788]: I1010 16:30:52.186355 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-3e16-account-create-xzt2v" event={"ID":"081f5e53-b6ea-41d5-ad7a-e1c11454af15","Type":"ContainerStarted","Data":"33de9e3c5a6f9508d58d58efab53f7b648d6364744af98da87fb74ab588ee04b"} Oct 10 16:30:53 crc kubenswrapper[4788]: I1010 16:30:53.035177 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-257v4"] Oct 10 16:30:53 crc kubenswrapper[4788]: I1010 16:30:53.051694 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-257v4"] Oct 10 16:30:53 crc kubenswrapper[4788]: I1010 16:30:53.234502 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:30:53 crc kubenswrapper[4788]: E1010 16:30:53.236190 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:30:53 crc kubenswrapper[4788]: I1010 16:30:53.596089 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-3e16-account-create-xzt2v" Oct 10 16:30:53 crc kubenswrapper[4788]: I1010 16:30:53.624429 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9drq\" (UniqueName: \"kubernetes.io/projected/081f5e53-b6ea-41d5-ad7a-e1c11454af15-kube-api-access-r9drq\") pod \"081f5e53-b6ea-41d5-ad7a-e1c11454af15\" (UID: \"081f5e53-b6ea-41d5-ad7a-e1c11454af15\") " Oct 10 16:30:53 crc kubenswrapper[4788]: I1010 16:30:53.632327 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/081f5e53-b6ea-41d5-ad7a-e1c11454af15-kube-api-access-r9drq" (OuterVolumeSpecName: "kube-api-access-r9drq") pod "081f5e53-b6ea-41d5-ad7a-e1c11454af15" (UID: "081f5e53-b6ea-41d5-ad7a-e1c11454af15"). InnerVolumeSpecName "kube-api-access-r9drq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:30:53 crc kubenswrapper[4788]: I1010 16:30:53.727850 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9drq\" (UniqueName: \"kubernetes.io/projected/081f5e53-b6ea-41d5-ad7a-e1c11454af15-kube-api-access-r9drq\") on node \"crc\" DevicePath \"\"" Oct 10 16:30:54 crc kubenswrapper[4788]: I1010 16:30:54.212935 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-3e16-account-create-xzt2v" event={"ID":"081f5e53-b6ea-41d5-ad7a-e1c11454af15","Type":"ContainerDied","Data":"33de9e3c5a6f9508d58d58efab53f7b648d6364744af98da87fb74ab588ee04b"} Oct 10 16:30:54 crc kubenswrapper[4788]: I1010 16:30:54.212978 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-3e16-account-create-xzt2v" Oct 10 16:30:54 crc kubenswrapper[4788]: I1010 16:30:54.213135 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33de9e3c5a6f9508d58d58efab53f7b648d6364744af98da87fb74ab588ee04b" Oct 10 16:30:54 crc kubenswrapper[4788]: I1010 16:30:54.254338 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8473942-7ab7-4e67-bc42-01e6c06d088f" path="/var/lib/kubelet/pods/a8473942-7ab7-4e67-bc42-01e6c06d088f/volumes" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.644292 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-pgqwd"] Oct 10 16:30:56 crc kubenswrapper[4788]: E1010 16:30:56.645819 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081f5e53-b6ea-41d5-ad7a-e1c11454af15" containerName="mariadb-account-create" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.645836 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="081f5e53-b6ea-41d5-ad7a-e1c11454af15" containerName="mariadb-account-create" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.646351 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="081f5e53-b6ea-41d5-ad7a-e1c11454af15" containerName="mariadb-account-create" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.647750 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.652773 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-zvprb" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.653115 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.685301 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-pgqwd"] Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.794824 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-job-config-data\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.794870 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgcnq\" (UniqueName: \"kubernetes.io/projected/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-kube-api-access-vgcnq\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.794943 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-combined-ca-bundle\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.795032 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-config-data\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.897311 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-config-data\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.897429 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-job-config-data\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.897452 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgcnq\" (UniqueName: \"kubernetes.io/projected/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-kube-api-access-vgcnq\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.897537 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-combined-ca-bundle\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.907809 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-config-data\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.909497 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-job-config-data\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.913353 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-combined-ca-bundle\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:56 crc kubenswrapper[4788]: I1010 16:30:56.925302 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgcnq\" (UniqueName: \"kubernetes.io/projected/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-kube-api-access-vgcnq\") pod \"manila-db-sync-pgqwd\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:57 crc kubenswrapper[4788]: I1010 16:30:57.016351 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-pgqwd" Oct 10 16:30:57 crc kubenswrapper[4788]: I1010 16:30:57.779473 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-pgqwd"] Oct 10 16:30:58 crc kubenswrapper[4788]: I1010 16:30:58.258842 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-pgqwd" event={"ID":"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081","Type":"ContainerStarted","Data":"baba9466c36ecf5240425aaf1b5b5e7ff180b493dae263fc905909533f81bb3f"} Oct 10 16:30:59 crc kubenswrapper[4788]: I1010 16:30:59.308247 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 16:31:07 crc kubenswrapper[4788]: I1010 16:31:07.356268 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-pgqwd" event={"ID":"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081","Type":"ContainerStarted","Data":"935e4bc58e2a8cc3ce8c8917424ab739ab5a1e82e23f2013ef1253f5545fca4d"} Oct 10 16:31:07 crc kubenswrapper[4788]: I1010 16:31:07.389492 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-pgqwd" podStartSLOduration=3.06208565 podStartE2EDuration="11.389477175s" podCreationTimestamp="2025-10-10 16:30:56 +0000 UTC" firstStartedPulling="2025-10-10 16:30:57.784945262 +0000 UTC m=+6360.234660810" lastFinishedPulling="2025-10-10 16:31:06.112336787 +0000 UTC m=+6368.562052335" observedRunningTime="2025-10-10 16:31:07.379850881 +0000 UTC m=+6369.829566429" watchObservedRunningTime="2025-10-10 16:31:07.389477175 +0000 UTC m=+6369.839192723" Oct 10 16:31:08 crc kubenswrapper[4788]: I1010 16:31:08.240406 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:31:08 crc kubenswrapper[4788]: E1010 16:31:08.240950 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:31:08 crc kubenswrapper[4788]: I1010 16:31:08.432567 4788 generic.go:334] "Generic (PLEG): container finished" podID="3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" containerID="935e4bc58e2a8cc3ce8c8917424ab739ab5a1e82e23f2013ef1253f5545fca4d" exitCode=0 Oct 10 16:31:08 crc kubenswrapper[4788]: I1010 16:31:08.432625 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-pgqwd" event={"ID":"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081","Type":"ContainerDied","Data":"935e4bc58e2a8cc3ce8c8917424ab739ab5a1e82e23f2013ef1253f5545fca4d"} Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.048500 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-pgqwd" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.156920 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgcnq\" (UniqueName: \"kubernetes.io/projected/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-kube-api-access-vgcnq\") pod \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.156980 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-config-data\") pod \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.157245 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-combined-ca-bundle\") pod \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.157323 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-job-config-data\") pod \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\" (UID: \"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081\") " Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.164336 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" (UID: "3bc6dd5e-8a7b-4dd9-b766-c23d089eb081"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.165711 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-kube-api-access-vgcnq" (OuterVolumeSpecName: "kube-api-access-vgcnq") pod "3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" (UID: "3bc6dd5e-8a7b-4dd9-b766-c23d089eb081"). InnerVolumeSpecName "kube-api-access-vgcnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.170794 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-config-data" (OuterVolumeSpecName: "config-data") pod "3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" (UID: "3bc6dd5e-8a7b-4dd9-b766-c23d089eb081"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.199279 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" (UID: "3bc6dd5e-8a7b-4dd9-b766-c23d089eb081"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.260884 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.260932 4788 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.260949 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgcnq\" (UniqueName: \"kubernetes.io/projected/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-kube-api-access-vgcnq\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.260964 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.462258 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-pgqwd" event={"ID":"3bc6dd5e-8a7b-4dd9-b766-c23d089eb081","Type":"ContainerDied","Data":"baba9466c36ecf5240425aaf1b5b5e7ff180b493dae263fc905909533f81bb3f"} Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.462308 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baba9466c36ecf5240425aaf1b5b5e7ff180b493dae263fc905909533f81bb3f" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.462362 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-pgqwd" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.819293 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 16:31:10 crc kubenswrapper[4788]: E1010 16:31:10.830703 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" containerName="manila-db-sync" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.830749 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" containerName="manila-db-sync" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.831068 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" containerName="manila-db-sync" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.832283 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.841389 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.841569 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.841787 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.844106 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-zvprb" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.850609 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.852947 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.856350 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.869872 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.885554 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.885616 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.885695 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj5br\" (UniqueName: \"kubernetes.io/projected/732e8571-c5ae-4750-a39a-b51da05c0385-kube-api-access-zj5br\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.885756 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-scripts\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.885853 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-config-data\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.885906 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/732e8571-c5ae-4750-a39a-b51da05c0385-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.915470 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.987172 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9949c5bf-d9nvn"] Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.988881 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991229 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991285 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3499fc18-c5b2-4a16-9db0-3f74ba613448-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991333 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-scripts\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991360 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-config\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991397 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991437 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kxnt\" (UniqueName: \"kubernetes.io/projected/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-kube-api-access-2kxnt\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991489 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991531 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-config-data\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991561 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-config-data\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991600 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/732e8571-c5ae-4750-a39a-b51da05c0385-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991649 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991680 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3499fc18-c5b2-4a16-9db0-3f74ba613448-ceph\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991713 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991744 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w2vs\" (UniqueName: \"kubernetes.io/projected/3499fc18-c5b2-4a16-9db0-3f74ba613448-kube-api-access-7w2vs\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991784 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-dns-svc\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991812 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991837 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-scripts\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991863 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3499fc18-c5b2-4a16-9db0-3f74ba613448-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.991932 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj5br\" (UniqueName: \"kubernetes.io/projected/732e8571-c5ae-4750-a39a-b51da05c0385-kube-api-access-zj5br\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.993495 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/732e8571-c5ae-4750-a39a-b51da05c0385-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:10 crc kubenswrapper[4788]: I1010 16:31:10.995039 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9949c5bf-d9nvn"] Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:10.999184 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-config-data\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.000878 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.013574 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.027616 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/732e8571-c5ae-4750-a39a-b51da05c0385-scripts\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.029893 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj5br\" (UniqueName: \"kubernetes.io/projected/732e8571-c5ae-4750-a39a-b51da05c0385-kube-api-access-zj5br\") pod \"manila-scheduler-0\" (UID: \"732e8571-c5ae-4750-a39a-b51da05c0385\") " pod="openstack/manila-scheduler-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094392 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094442 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3499fc18-c5b2-4a16-9db0-3f74ba613448-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094488 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-config\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094520 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094547 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kxnt\" (UniqueName: \"kubernetes.io/projected/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-kube-api-access-2kxnt\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094580 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094607 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-config-data\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094649 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094667 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3499fc18-c5b2-4a16-9db0-3f74ba613448-ceph\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094693 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w2vs\" (UniqueName: \"kubernetes.io/projected/3499fc18-c5b2-4a16-9db0-3f74ba613448-kube-api-access-7w2vs\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094720 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-dns-svc\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094745 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-scripts\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094762 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3499fc18-c5b2-4a16-9db0-3f74ba613448-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.094879 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3499fc18-c5b2-4a16-9db0-3f74ba613448-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.095529 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3499fc18-c5b2-4a16-9db0-3f74ba613448-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.096131 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.097028 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-config\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.097565 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-dns-svc\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.101500 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.103409 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.108287 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.109319 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-config-data\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.109587 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.110536 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-scripts\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.119257 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.119886 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3499fc18-c5b2-4a16-9db0-3f74ba613448-ceph\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.122284 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3499fc18-c5b2-4a16-9db0-3f74ba613448-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.123621 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w2vs\" (UniqueName: \"kubernetes.io/projected/3499fc18-c5b2-4a16-9db0-3f74ba613448-kube-api-access-7w2vs\") pod \"manila-share-share1-0\" (UID: \"3499fc18-c5b2-4a16-9db0-3f74ba613448\") " pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.130082 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kxnt\" (UniqueName: \"kubernetes.io/projected/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-kube-api-access-2kxnt\") pod \"dnsmasq-dns-5c9949c5bf-d9nvn\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.134127 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.181234 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.215653 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.282510 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.297619 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-scripts\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.297709 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-config-data-custom\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.298134 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9n74\" (UniqueName: \"kubernetes.io/projected/f6ed40a0-c7d7-4213-82f5-96fc17015781-kube-api-access-p9n74\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.298221 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6ed40a0-c7d7-4213-82f5-96fc17015781-logs\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.298296 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ed40a0-c7d7-4213-82f5-96fc17015781-etc-machine-id\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.300325 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.300719 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-config-data\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.401930 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.402001 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-config-data\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.402039 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-scripts\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.402127 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-config-data-custom\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.402187 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9n74\" (UniqueName: \"kubernetes.io/projected/f6ed40a0-c7d7-4213-82f5-96fc17015781-kube-api-access-p9n74\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.402219 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6ed40a0-c7d7-4213-82f5-96fc17015781-logs\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.402290 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ed40a0-c7d7-4213-82f5-96fc17015781-etc-machine-id\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.404995 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f6ed40a0-c7d7-4213-82f5-96fc17015781-etc-machine-id\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.405795 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6ed40a0-c7d7-4213-82f5-96fc17015781-logs\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.409834 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-config-data-custom\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.411087 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-config-data\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.411430 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-scripts\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.422691 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ed40a0-c7d7-4213-82f5-96fc17015781-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.428725 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9n74\" (UniqueName: \"kubernetes.io/projected/f6ed40a0-c7d7-4213-82f5-96fc17015781-kube-api-access-p9n74\") pod \"manila-api-0\" (UID: \"f6ed40a0-c7d7-4213-82f5-96fc17015781\") " pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.592635 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 10 16:31:11 crc kubenswrapper[4788]: I1010 16:31:11.796630 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 10 16:31:11 crc kubenswrapper[4788]: W1010 16:31:11.857887 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod732e8571_c5ae_4750_a39a_b51da05c0385.slice/crio-4220196b3b0439546b03af7eb52f3f88e4fae94747be53e6cb5e4d29d11545e1 WatchSource:0}: Error finding container 4220196b3b0439546b03af7eb52f3f88e4fae94747be53e6cb5e4d29d11545e1: Status 404 returned error can't find the container with id 4220196b3b0439546b03af7eb52f3f88e4fae94747be53e6cb5e4d29d11545e1 Oct 10 16:31:12 crc kubenswrapper[4788]: I1010 16:31:12.074625 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 10 16:31:12 crc kubenswrapper[4788]: W1010 16:31:12.078628 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3499fc18_c5b2_4a16_9db0_3f74ba613448.slice/crio-66496b5708fd16a2526919536872ce2d60baac2fe7bf073d0f2f84712c16c4f6 WatchSource:0}: Error finding container 66496b5708fd16a2526919536872ce2d60baac2fe7bf073d0f2f84712c16c4f6: Status 404 returned error can't find the container with id 66496b5708fd16a2526919536872ce2d60baac2fe7bf073d0f2f84712c16c4f6 Oct 10 16:31:12 crc kubenswrapper[4788]: I1010 16:31:12.116687 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9949c5bf-d9nvn"] Oct 10 16:31:12 crc kubenswrapper[4788]: I1010 16:31:12.485143 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 10 16:31:12 crc kubenswrapper[4788]: I1010 16:31:12.521820 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"732e8571-c5ae-4750-a39a-b51da05c0385","Type":"ContainerStarted","Data":"4220196b3b0439546b03af7eb52f3f88e4fae94747be53e6cb5e4d29d11545e1"} Oct 10 16:31:12 crc kubenswrapper[4788]: I1010 16:31:12.524322 4788 generic.go:334] "Generic (PLEG): container finished" podID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerID="28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31" exitCode=0 Oct 10 16:31:12 crc kubenswrapper[4788]: I1010 16:31:12.524360 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" event={"ID":"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc","Type":"ContainerDied","Data":"28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31"} Oct 10 16:31:12 crc kubenswrapper[4788]: I1010 16:31:12.524392 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" event={"ID":"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc","Type":"ContainerStarted","Data":"de815f3e11cf3d4533bf29fb156122121c8abd15d068aa4c93dcb25174ed9502"} Oct 10 16:31:12 crc kubenswrapper[4788]: I1010 16:31:12.526063 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3499fc18-c5b2-4a16-9db0-3f74ba613448","Type":"ContainerStarted","Data":"66496b5708fd16a2526919536872ce2d60baac2fe7bf073d0f2f84712c16c4f6"} Oct 10 16:31:12 crc kubenswrapper[4788]: W1010 16:31:12.547823 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6ed40a0_c7d7_4213_82f5_96fc17015781.slice/crio-dd100d51b671ff6fff83ac260b1c949ca34e237dae4ab58d0d0d2b79064cd993 WatchSource:0}: Error finding container dd100d51b671ff6fff83ac260b1c949ca34e237dae4ab58d0d0d2b79064cd993: Status 404 returned error can't find the container with id dd100d51b671ff6fff83ac260b1c949ca34e237dae4ab58d0d0d2b79064cd993 Oct 10 16:31:13 crc kubenswrapper[4788]: I1010 16:31:13.584395 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" event={"ID":"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc","Type":"ContainerStarted","Data":"63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7"} Oct 10 16:31:13 crc kubenswrapper[4788]: I1010 16:31:13.586238 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:13 crc kubenswrapper[4788]: I1010 16:31:13.627466 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" podStartSLOduration=3.6274492560000002 podStartE2EDuration="3.627449256s" podCreationTimestamp="2025-10-10 16:31:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:31:13.619397373 +0000 UTC m=+6376.069112921" watchObservedRunningTime="2025-10-10 16:31:13.627449256 +0000 UTC m=+6376.077164804" Oct 10 16:31:13 crc kubenswrapper[4788]: I1010 16:31:13.658648 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6ed40a0-c7d7-4213-82f5-96fc17015781","Type":"ContainerStarted","Data":"1bda651b57c409d0021401235251d7ba0d6eb08a4952c4cd800a55878544426b"} Oct 10 16:31:13 crc kubenswrapper[4788]: I1010 16:31:13.658696 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6ed40a0-c7d7-4213-82f5-96fc17015781","Type":"ContainerStarted","Data":"dd100d51b671ff6fff83ac260b1c949ca34e237dae4ab58d0d0d2b79064cd993"} Oct 10 16:31:14 crc kubenswrapper[4788]: I1010 16:31:14.686970 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"732e8571-c5ae-4750-a39a-b51da05c0385","Type":"ContainerStarted","Data":"2eca37ed3a85a9a8888b07c5a805417a9cdbd1049dbafe85f7e57c93e9b6cf61"} Oct 10 16:31:14 crc kubenswrapper[4788]: I1010 16:31:14.689976 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"732e8571-c5ae-4750-a39a-b51da05c0385","Type":"ContainerStarted","Data":"e957849bf0be6b1dee1c8961b654f76b5d950145bc750bd1f613030d56eb504f"} Oct 10 16:31:14 crc kubenswrapper[4788]: I1010 16:31:14.695049 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f6ed40a0-c7d7-4213-82f5-96fc17015781","Type":"ContainerStarted","Data":"b7a7dbf604a387724a2d91d27a4b1b2f3c2fa26c7a2fcfa7680ae4f499077c67"} Oct 10 16:31:14 crc kubenswrapper[4788]: I1010 16:31:14.695560 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 10 16:31:14 crc kubenswrapper[4788]: I1010 16:31:14.713890 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.358320491 podStartE2EDuration="4.713870189s" podCreationTimestamp="2025-10-10 16:31:10 +0000 UTC" firstStartedPulling="2025-10-10 16:31:11.875475992 +0000 UTC m=+6374.325191550" lastFinishedPulling="2025-10-10 16:31:13.2310257 +0000 UTC m=+6375.680741248" observedRunningTime="2025-10-10 16:31:14.707534732 +0000 UTC m=+6377.157250300" watchObservedRunningTime="2025-10-10 16:31:14.713870189 +0000 UTC m=+6377.163585737" Oct 10 16:31:14 crc kubenswrapper[4788]: I1010 16:31:14.739188 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.7391442870000002 podStartE2EDuration="3.739144287s" podCreationTimestamp="2025-10-10 16:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:31:14.729001829 +0000 UTC m=+6377.178717367" watchObservedRunningTime="2025-10-10 16:31:14.739144287 +0000 UTC m=+6377.188859835" Oct 10 16:31:17 crc kubenswrapper[4788]: I1010 16:31:17.859971 4788 scope.go:117] "RemoveContainer" containerID="5d20d1006436fbf2b65d3bd508ba74c1355a9279ab8fe1284ee857ff1ec7cd7a" Oct 10 16:31:17 crc kubenswrapper[4788]: I1010 16:31:17.907853 4788 scope.go:117] "RemoveContainer" containerID="2aeb26981bc33b85edc78f4b183b3255c039e4977d876d4c4d00fbc1d3f25b1b" Oct 10 16:31:17 crc kubenswrapper[4788]: I1010 16:31:17.952212 4788 scope.go:117] "RemoveContainer" containerID="c3d1eb399feb77ea0c3badde3438508613091fcaac460dbea92e86d2d067e614" Oct 10 16:31:19 crc kubenswrapper[4788]: I1010 16:31:19.234577 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:31:19 crc kubenswrapper[4788]: E1010 16:31:19.235347 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:31:21 crc kubenswrapper[4788]: I1010 16:31:21.182443 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 10 16:31:21 crc kubenswrapper[4788]: I1010 16:31:21.285127 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:31:21 crc kubenswrapper[4788]: I1010 16:31:21.347187 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b757d79cc-8ltdz"] Oct 10 16:31:21 crc kubenswrapper[4788]: I1010 16:31:21.347450 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" podUID="e2664a06-b38b-49b9-8651-75098ac9d1f0" containerName="dnsmasq-dns" containerID="cri-o://4d6b1a58dce6ea3ad07e4b51485fe47f33336351e659566710c66494646bff72" gracePeriod=10 Oct 10 16:31:21 crc kubenswrapper[4788]: I1010 16:31:21.770398 4788 generic.go:334] "Generic (PLEG): container finished" podID="e2664a06-b38b-49b9-8651-75098ac9d1f0" containerID="4d6b1a58dce6ea3ad07e4b51485fe47f33336351e659566710c66494646bff72" exitCode=0 Oct 10 16:31:21 crc kubenswrapper[4788]: I1010 16:31:21.770807 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" event={"ID":"e2664a06-b38b-49b9-8651-75098ac9d1f0","Type":"ContainerDied","Data":"4d6b1a58dce6ea3ad07e4b51485fe47f33336351e659566710c66494646bff72"} Oct 10 16:31:21 crc kubenswrapper[4788]: I1010 16:31:21.944434 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.112419 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5whx\" (UniqueName: \"kubernetes.io/projected/e2664a06-b38b-49b9-8651-75098ac9d1f0-kube-api-access-l5whx\") pod \"e2664a06-b38b-49b9-8651-75098ac9d1f0\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.112506 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-sb\") pod \"e2664a06-b38b-49b9-8651-75098ac9d1f0\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.112592 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-dns-svc\") pod \"e2664a06-b38b-49b9-8651-75098ac9d1f0\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.112640 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-nb\") pod \"e2664a06-b38b-49b9-8651-75098ac9d1f0\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.112767 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-config\") pod \"e2664a06-b38b-49b9-8651-75098ac9d1f0\" (UID: \"e2664a06-b38b-49b9-8651-75098ac9d1f0\") " Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.122448 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2664a06-b38b-49b9-8651-75098ac9d1f0-kube-api-access-l5whx" (OuterVolumeSpecName: "kube-api-access-l5whx") pod "e2664a06-b38b-49b9-8651-75098ac9d1f0" (UID: "e2664a06-b38b-49b9-8651-75098ac9d1f0"). InnerVolumeSpecName "kube-api-access-l5whx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.167838 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e2664a06-b38b-49b9-8651-75098ac9d1f0" (UID: "e2664a06-b38b-49b9-8651-75098ac9d1f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.171318 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-config" (OuterVolumeSpecName: "config") pod "e2664a06-b38b-49b9-8651-75098ac9d1f0" (UID: "e2664a06-b38b-49b9-8651-75098ac9d1f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.176900 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e2664a06-b38b-49b9-8651-75098ac9d1f0" (UID: "e2664a06-b38b-49b9-8651-75098ac9d1f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.184350 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e2664a06-b38b-49b9-8651-75098ac9d1f0" (UID: "e2664a06-b38b-49b9-8651-75098ac9d1f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.215680 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5whx\" (UniqueName: \"kubernetes.io/projected/e2664a06-b38b-49b9-8651-75098ac9d1f0-kube-api-access-l5whx\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.215726 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.215742 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.215754 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.215769 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2664a06-b38b-49b9-8651-75098ac9d1f0-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.785107 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" event={"ID":"e2664a06-b38b-49b9-8651-75098ac9d1f0","Type":"ContainerDied","Data":"53a88bd84d1bc7224169f11c60ef2913d00978526940abe206fbadd315782897"} Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.785177 4788 scope.go:117] "RemoveContainer" containerID="4d6b1a58dce6ea3ad07e4b51485fe47f33336351e659566710c66494646bff72" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.785312 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b757d79cc-8ltdz" Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.818736 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b757d79cc-8ltdz"] Oct 10 16:31:22 crc kubenswrapper[4788]: I1010 16:31:22.832270 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b757d79cc-8ltdz"] Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.248432 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2664a06-b38b-49b9-8651-75098ac9d1f0" path="/var/lib/kubelet/pods/e2664a06-b38b-49b9-8651-75098ac9d1f0/volumes" Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.562652 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.562905 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="ceilometer-central-agent" containerID="cri-o://7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c" gracePeriod=30 Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.563396 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="proxy-httpd" containerID="cri-o://0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b" gracePeriod=30 Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.563475 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="ceilometer-notification-agent" containerID="cri-o://4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf" gracePeriod=30 Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.563510 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="sg-core" containerID="cri-o://ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329" gracePeriod=30 Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.646565 4788 scope.go:117] "RemoveContainer" containerID="7fb9f350e115b134ad6783b35fe697699d2d135e30c15516b9bdf4f510aa953b" Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.854006 4788 generic.go:334] "Generic (PLEG): container finished" podID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerID="0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b" exitCode=0 Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.854327 4788 generic.go:334] "Generic (PLEG): container finished" podID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerID="ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329" exitCode=2 Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.854089 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerDied","Data":"0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b"} Oct 10 16:31:24 crc kubenswrapper[4788]: I1010 16:31:24.854396 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerDied","Data":"ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329"} Oct 10 16:31:25 crc kubenswrapper[4788]: I1010 16:31:25.869977 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3499fc18-c5b2-4a16-9db0-3f74ba613448","Type":"ContainerStarted","Data":"8f5b91da63c65c3e5b057437c8fb8f9c8e76ad1ff9fc1e20981d749bbef52e5c"} Oct 10 16:31:25 crc kubenswrapper[4788]: I1010 16:31:25.870619 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3499fc18-c5b2-4a16-9db0-3f74ba613448","Type":"ContainerStarted","Data":"18024d4db63cb3e43d3eca27b101a408b6785a5db38352b3aa2dc704e77e3426"} Oct 10 16:31:25 crc kubenswrapper[4788]: I1010 16:31:25.877218 4788 generic.go:334] "Generic (PLEG): container finished" podID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerID="7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c" exitCode=0 Oct 10 16:31:25 crc kubenswrapper[4788]: I1010 16:31:25.877269 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerDied","Data":"7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c"} Oct 10 16:31:25 crc kubenswrapper[4788]: I1010 16:31:25.891693 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.246890097 podStartE2EDuration="15.891674707s" podCreationTimestamp="2025-10-10 16:31:10 +0000 UTC" firstStartedPulling="2025-10-10 16:31:12.080590047 +0000 UTC m=+6374.530305595" lastFinishedPulling="2025-10-10 16:31:24.725374657 +0000 UTC m=+6387.175090205" observedRunningTime="2025-10-10 16:31:25.891029971 +0000 UTC m=+6388.340745519" watchObservedRunningTime="2025-10-10 16:31:25.891674707 +0000 UTC m=+6388.341390255" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.304335 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.1.142:3000/\": dial tcp 10.217.1.142:3000: connect: connection refused" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.761756 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.909376 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-scripts\") pod \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.909446 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-sg-core-conf-yaml\") pod \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.909497 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z22w9\" (UniqueName: \"kubernetes.io/projected/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-kube-api-access-z22w9\") pod \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.909546 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-combined-ca-bundle\") pod \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.909773 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-config-data\") pod \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.909834 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-run-httpd\") pod \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.909906 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-log-httpd\") pod \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\" (UID: \"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5\") " Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.911276 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" (UID: "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.912599 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" (UID: "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.921753 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-scripts" (OuterVolumeSpecName: "scripts") pod "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" (UID: "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.935134 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-kube-api-access-z22w9" (OuterVolumeSpecName: "kube-api-access-z22w9") pod "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" (UID: "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5"). InnerVolumeSpecName "kube-api-access-z22w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.944735 4788 generic.go:334] "Generic (PLEG): container finished" podID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerID="4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf" exitCode=0 Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.944783 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerDied","Data":"4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf"} Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.944812 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5","Type":"ContainerDied","Data":"96e5a308e2c8d07b186678227c5000c0cf29be920b713f929426fbb23e34b8e5"} Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.944831 4788 scope.go:117] "RemoveContainer" containerID="0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.944990 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:31:29 crc kubenswrapper[4788]: I1010 16:31:29.966370 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" (UID: "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.012476 4788 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-scripts\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.012514 4788 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.012527 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z22w9\" (UniqueName: \"kubernetes.io/projected/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-kube-api-access-z22w9\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.012537 4788 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.012548 4788 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.029734 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" (UID: "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.047080 4788 scope.go:117] "RemoveContainer" containerID="ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.071056 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-config-data" (OuterVolumeSpecName: "config-data") pod "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" (UID: "5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.075796 4788 scope.go:117] "RemoveContainer" containerID="4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.096478 4788 scope.go:117] "RemoveContainer" containerID="7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.114238 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.114266 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.115873 4788 scope.go:117] "RemoveContainer" containerID="0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.116200 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b\": container with ID starting with 0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b not found: ID does not exist" containerID="0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.116242 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b"} err="failed to get container status \"0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b\": rpc error: code = NotFound desc = could not find container \"0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b\": container with ID starting with 0fdebd102f636ceeb8e12d877a4d469da24910b85d89fa0253d4a9f9984d694b not found: ID does not exist" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.116266 4788 scope.go:117] "RemoveContainer" containerID="ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.116856 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329\": container with ID starting with ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329 not found: ID does not exist" containerID="ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.116877 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329"} err="failed to get container status \"ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329\": rpc error: code = NotFound desc = could not find container \"ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329\": container with ID starting with ca84e54cc2ff2267d5dada915653d7880233ad9498849f3213831f27a93c2329 not found: ID does not exist" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.116890 4788 scope.go:117] "RemoveContainer" containerID="4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.117106 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf\": container with ID starting with 4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf not found: ID does not exist" containerID="4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.117153 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf"} err="failed to get container status \"4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf\": rpc error: code = NotFound desc = could not find container \"4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf\": container with ID starting with 4a75fa7cb9cde43f1b23cc34ed3dc191c42b73ec3cf7e9cb3e219c4ce3b775bf not found: ID does not exist" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.117168 4788 scope.go:117] "RemoveContainer" containerID="7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.117407 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c\": container with ID starting with 7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c not found: ID does not exist" containerID="7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.117427 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c"} err="failed to get container status \"7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c\": rpc error: code = NotFound desc = could not find container \"7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c\": container with ID starting with 7b920dc070881c51a7523329c86f676f468a8e53f5feac3f8bfaa9e11717175c not found: ID does not exist" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.233881 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.234847 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.309711 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.334571 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.348553 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.349030 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2664a06-b38b-49b9-8651-75098ac9d1f0" containerName="init" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349050 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2664a06-b38b-49b9-8651-75098ac9d1f0" containerName="init" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.349079 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2664a06-b38b-49b9-8651-75098ac9d1f0" containerName="dnsmasq-dns" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349085 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2664a06-b38b-49b9-8651-75098ac9d1f0" containerName="dnsmasq-dns" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.349103 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="proxy-httpd" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349110 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="proxy-httpd" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.349125 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="ceilometer-central-agent" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349130 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="ceilometer-central-agent" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.349171 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="sg-core" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349178 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="sg-core" Oct 10 16:31:30 crc kubenswrapper[4788]: E1010 16:31:30.349211 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="ceilometer-notification-agent" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349218 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="ceilometer-notification-agent" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349436 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="sg-core" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349451 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="ceilometer-central-agent" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349463 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="ceilometer-notification-agent" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349475 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" containerName="proxy-httpd" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.349494 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2664a06-b38b-49b9-8651-75098ac9d1f0" containerName="dnsmasq-dns" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.351576 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.355851 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.356105 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.361888 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.522737 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d5dfc2e-d845-48b9-8741-d9972bbab228-log-httpd\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.522951 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-scripts\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.523033 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d5dfc2e-d845-48b9-8741-d9972bbab228-run-httpd\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.523207 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.523316 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsvwq\" (UniqueName: \"kubernetes.io/projected/6d5dfc2e-d845-48b9-8741-d9972bbab228-kube-api-access-zsvwq\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.523938 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.524027 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-config-data\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.626198 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsvwq\" (UniqueName: \"kubernetes.io/projected/6d5dfc2e-d845-48b9-8741-d9972bbab228-kube-api-access-zsvwq\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.626414 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.626472 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-config-data\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.626558 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d5dfc2e-d845-48b9-8741-d9972bbab228-log-httpd\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.626661 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-scripts\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.626758 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d5dfc2e-d845-48b9-8741-d9972bbab228-run-httpd\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.626817 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.627597 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d5dfc2e-d845-48b9-8741-d9972bbab228-log-httpd\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.628061 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d5dfc2e-d845-48b9-8741-d9972bbab228-run-httpd\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.633215 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-config-data\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.633229 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.633539 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.633813 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d5dfc2e-d845-48b9-8741-d9972bbab228-scripts\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.654656 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsvwq\" (UniqueName: \"kubernetes.io/projected/6d5dfc2e-d845-48b9-8741-d9972bbab228-kube-api-access-zsvwq\") pod \"ceilometer-0\" (UID: \"6d5dfc2e-d845-48b9-8741-d9972bbab228\") " pod="openstack/ceilometer-0" Oct 10 16:31:30 crc kubenswrapper[4788]: I1010 16:31:30.683197 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 10 16:31:31 crc kubenswrapper[4788]: I1010 16:31:31.216499 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 10 16:31:31 crc kubenswrapper[4788]: I1010 16:31:31.230828 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 10 16:31:31 crc kubenswrapper[4788]: W1010 16:31:31.241072 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d5dfc2e_d845_48b9_8741_d9972bbab228.slice/crio-eed27a62f991d01242bb17c3676c7960026658c3bf02c4ae1f5484624a4cf672 WatchSource:0}: Error finding container eed27a62f991d01242bb17c3676c7960026658c3bf02c4ae1f5484624a4cf672: Status 404 returned error can't find the container with id eed27a62f991d01242bb17c3676c7960026658c3bf02c4ae1f5484624a4cf672 Oct 10 16:31:32 crc kubenswrapper[4788]: I1010 16:31:32.006343 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d5dfc2e-d845-48b9-8741-d9972bbab228","Type":"ContainerStarted","Data":"eed27a62f991d01242bb17c3676c7960026658c3bf02c4ae1f5484624a4cf672"} Oct 10 16:31:32 crc kubenswrapper[4788]: I1010 16:31:32.248837 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5" path="/var/lib/kubelet/pods/5b9cb3b0-f31b-40b8-9a2a-c9fc4553aed5/volumes" Oct 10 16:31:32 crc kubenswrapper[4788]: I1010 16:31:32.983913 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 10 16:31:33 crc kubenswrapper[4788]: I1010 16:31:33.030516 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d5dfc2e-d845-48b9-8741-d9972bbab228","Type":"ContainerStarted","Data":"5cc5b5597944ab3c3f70bc96120904a40b7d06a2feb5ca5d898542d60253e8d3"} Oct 10 16:31:33 crc kubenswrapper[4788]: I1010 16:31:33.114794 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 10 16:31:34 crc kubenswrapper[4788]: I1010 16:31:34.043288 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d5dfc2e-d845-48b9-8741-d9972bbab228","Type":"ContainerStarted","Data":"17339186e0a92dac6c720350e25f97827915086ac2345e6aa8a6ff3da15a122e"} Oct 10 16:31:35 crc kubenswrapper[4788]: I1010 16:31:35.053666 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d5dfc2e-d845-48b9-8741-d9972bbab228","Type":"ContainerStarted","Data":"3d35e79869e884b396c2fcb51b554f4d1c712969d2c4f56ca5dc48e51dbb4ce0"} Oct 10 16:31:39 crc kubenswrapper[4788]: I1010 16:31:39.109416 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d5dfc2e-d845-48b9-8741-d9972bbab228","Type":"ContainerStarted","Data":"7bf014d71b0b8002047bd80d6fc7528800919922e35fd5f1bb46958f92231b2f"} Oct 10 16:31:39 crc kubenswrapper[4788]: I1010 16:31:39.111289 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 10 16:31:39 crc kubenswrapper[4788]: I1010 16:31:39.157479 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.628548011 podStartE2EDuration="9.157447783s" podCreationTimestamp="2025-10-10 16:31:30 +0000 UTC" firstStartedPulling="2025-10-10 16:31:31.244215762 +0000 UTC m=+6393.693931310" lastFinishedPulling="2025-10-10 16:31:37.773115524 +0000 UTC m=+6400.222831082" observedRunningTime="2025-10-10 16:31:39.149101832 +0000 UTC m=+6401.598817490" watchObservedRunningTime="2025-10-10 16:31:39.157447783 +0000 UTC m=+6401.607163371" Oct 10 16:31:42 crc kubenswrapper[4788]: I1010 16:31:42.962885 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 10 16:31:43 crc kubenswrapper[4788]: I1010 16:31:43.257208 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:31:43 crc kubenswrapper[4788]: E1010 16:31:43.257671 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:31:46 crc kubenswrapper[4788]: E1010 16:31:46.134541 4788 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.129:40692->38.102.83.129:41727: write tcp 38.102.83.129:40692->38.102.83.129:41727: write: broken pipe Oct 10 16:31:58 crc kubenswrapper[4788]: I1010 16:31:58.245782 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:31:58 crc kubenswrapper[4788]: E1010 16:31:58.247272 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:32:00 crc kubenswrapper[4788]: I1010 16:32:00.702055 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 10 16:32:13 crc kubenswrapper[4788]: I1010 16:32:13.236411 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:32:13 crc kubenswrapper[4788]: E1010 16:32:13.237821 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:32:26 crc kubenswrapper[4788]: I1010 16:32:26.234366 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:32:26 crc kubenswrapper[4788]: E1010 16:32:26.235131 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.394770 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c94656c95-77nqv"] Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.398452 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.401755 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.404061 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c94656c95-77nqv"] Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.492603 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-openstack-cell1\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.492697 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-nb\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.492800 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-sb\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.492866 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4bcv\" (UniqueName: \"kubernetes.io/projected/28d8736e-200c-4765-b78a-4a3913114f72-kube-api-access-l4bcv\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.492910 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-config\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.492957 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-dns-svc\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.594860 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-sb\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.594933 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4bcv\" (UniqueName: \"kubernetes.io/projected/28d8736e-200c-4765-b78a-4a3913114f72-kube-api-access-l4bcv\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.594970 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-config\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.595003 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-dns-svc\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.595173 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-openstack-cell1\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.595207 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-nb\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.595852 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-config\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.595917 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-dns-svc\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.596114 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-nb\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.596347 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-openstack-cell1\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.596412 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-sb\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.615775 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4bcv\" (UniqueName: \"kubernetes.io/projected/28d8736e-200c-4765-b78a-4a3913114f72-kube-api-access-l4bcv\") pod \"dnsmasq-dns-5c94656c95-77nqv\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:30 crc kubenswrapper[4788]: I1010 16:32:30.719329 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:31 crc kubenswrapper[4788]: I1010 16:32:31.407595 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c94656c95-77nqv"] Oct 10 16:32:31 crc kubenswrapper[4788]: I1010 16:32:31.738918 4788 generic.go:334] "Generic (PLEG): container finished" podID="28d8736e-200c-4765-b78a-4a3913114f72" containerID="831743d71cc6a393fcb1d57c687e875a6a662d3ef468cdebaf209fa48e78486f" exitCode=0 Oct 10 16:32:31 crc kubenswrapper[4788]: I1010 16:32:31.739191 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" event={"ID":"28d8736e-200c-4765-b78a-4a3913114f72","Type":"ContainerDied","Data":"831743d71cc6a393fcb1d57c687e875a6a662d3ef468cdebaf209fa48e78486f"} Oct 10 16:32:31 crc kubenswrapper[4788]: I1010 16:32:31.739215 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" event={"ID":"28d8736e-200c-4765-b78a-4a3913114f72","Type":"ContainerStarted","Data":"c6256cbb50ecd6904d50b47522d3ab744dd3d9bc18d0f12d3ce9784b80ce01b9"} Oct 10 16:32:32 crc kubenswrapper[4788]: I1010 16:32:32.749407 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" event={"ID":"28d8736e-200c-4765-b78a-4a3913114f72","Type":"ContainerStarted","Data":"dbca9f587ea5d14ae555bebda043de8949f825999129119d895fcb47c0f9308b"} Oct 10 16:32:32 crc kubenswrapper[4788]: I1010 16:32:32.749698 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:32 crc kubenswrapper[4788]: I1010 16:32:32.777843 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" podStartSLOduration=2.777823037 podStartE2EDuration="2.777823037s" podCreationTimestamp="2025-10-10 16:32:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:32:32.769640611 +0000 UTC m=+6455.219356159" watchObservedRunningTime="2025-10-10 16:32:32.777823037 +0000 UTC m=+6455.227538585" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.234322 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:32:40 crc kubenswrapper[4788]: E1010 16:32:40.235890 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.721431 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.805374 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9949c5bf-d9nvn"] Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.806156 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" podUID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerName="dnsmasq-dns" containerID="cri-o://63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7" gracePeriod=10 Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.952202 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b74f9b95-7wj79"] Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.954272 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.965535 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b74f9b95-7wj79"] Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.969785 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-openstack-cell1\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.969856 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-config\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.970016 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krmk7\" (UniqueName: \"kubernetes.io/projected/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-kube-api-access-krmk7\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.970085 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-dns-svc\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.970162 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-ovsdbserver-nb\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:40 crc kubenswrapper[4788]: I1010 16:32:40.970218 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-ovsdbserver-sb\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.071614 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krmk7\" (UniqueName: \"kubernetes.io/projected/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-kube-api-access-krmk7\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.071692 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-dns-svc\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.071730 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-ovsdbserver-nb\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.071767 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-ovsdbserver-sb\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.071824 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-openstack-cell1\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.071847 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-config\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.072853 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-dns-svc\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.072994 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-ovsdbserver-sb\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.073400 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-config\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.073643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-openstack-cell1\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.073775 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-ovsdbserver-nb\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.091072 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krmk7\" (UniqueName: \"kubernetes.io/projected/3c0ee6c9-eb8b-4054-a552-fccd55b68cfc-kube-api-access-krmk7\") pod \"dnsmasq-dns-84b74f9b95-7wj79\" (UID: \"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc\") " pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.309116 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.479484 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.589649 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kxnt\" (UniqueName: \"kubernetes.io/projected/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-kube-api-access-2kxnt\") pod \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.591596 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-config\") pod \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.591829 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-dns-svc\") pod \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.591880 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-nb\") pod \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.591936 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-sb\") pod \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\" (UID: \"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc\") " Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.632325 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-kube-api-access-2kxnt" (OuterVolumeSpecName: "kube-api-access-2kxnt") pod "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" (UID: "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc"). InnerVolumeSpecName "kube-api-access-2kxnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.668191 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-config" (OuterVolumeSpecName: "config") pod "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" (UID: "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.681545 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" (UID: "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.702933 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" (UID: "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.711376 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" (UID: "9c5ab940-e2a4-4024-b4aa-172ee2c6eafc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.721956 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.722010 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.722033 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.722048 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kxnt\" (UniqueName: \"kubernetes.io/projected/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-kube-api-access-2kxnt\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.722061 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.848446 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b74f9b95-7wj79"] Oct 10 16:32:41 crc kubenswrapper[4788]: W1010 16:32:41.857114 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c0ee6c9_eb8b_4054_a552_fccd55b68cfc.slice/crio-895c00276a579c667ba654b55d426818a768515e1359295c63b2e825f0cc11a2 WatchSource:0}: Error finding container 895c00276a579c667ba654b55d426818a768515e1359295c63b2e825f0cc11a2: Status 404 returned error can't find the container with id 895c00276a579c667ba654b55d426818a768515e1359295c63b2e825f0cc11a2 Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.874848 4788 generic.go:334] "Generic (PLEG): container finished" podID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerID="63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7" exitCode=0 Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.874922 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.874916 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" event={"ID":"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc","Type":"ContainerDied","Data":"63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7"} Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.875186 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" event={"ID":"9c5ab940-e2a4-4024-b4aa-172ee2c6eafc","Type":"ContainerDied","Data":"de815f3e11cf3d4533bf29fb156122121c8abd15d068aa4c93dcb25174ed9502"} Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.875222 4788 scope.go:117] "RemoveContainer" containerID="63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.937987 4788 scope.go:117] "RemoveContainer" containerID="28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31" Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.984303 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9949c5bf-d9nvn"] Oct 10 16:32:41 crc kubenswrapper[4788]: I1010 16:32:41.996388 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9949c5bf-d9nvn"] Oct 10 16:32:42 crc kubenswrapper[4788]: I1010 16:32:42.153205 4788 scope.go:117] "RemoveContainer" containerID="63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7" Oct 10 16:32:42 crc kubenswrapper[4788]: E1010 16:32:42.154592 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7\": container with ID starting with 63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7 not found: ID does not exist" containerID="63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7" Oct 10 16:32:42 crc kubenswrapper[4788]: I1010 16:32:42.154700 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7"} err="failed to get container status \"63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7\": rpc error: code = NotFound desc = could not find container \"63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7\": container with ID starting with 63ed0ebee5291540319ca9a9281b8a912bc27c449df48cd8944fa5056ce36ac7 not found: ID does not exist" Oct 10 16:32:42 crc kubenswrapper[4788]: I1010 16:32:42.154807 4788 scope.go:117] "RemoveContainer" containerID="28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31" Oct 10 16:32:42 crc kubenswrapper[4788]: E1010 16:32:42.155696 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31\": container with ID starting with 28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31 not found: ID does not exist" containerID="28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31" Oct 10 16:32:42 crc kubenswrapper[4788]: I1010 16:32:42.155810 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31"} err="failed to get container status \"28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31\": rpc error: code = NotFound desc = could not find container \"28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31\": container with ID starting with 28070d20a4170fded6048d199fd2fa5839c640bed0feb2dad165728f5e641d31 not found: ID does not exist" Oct 10 16:32:42 crc kubenswrapper[4788]: I1010 16:32:42.247519 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" path="/var/lib/kubelet/pods/9c5ab940-e2a4-4024-b4aa-172ee2c6eafc/volumes" Oct 10 16:32:42 crc kubenswrapper[4788]: I1010 16:32:42.887484 4788 generic.go:334] "Generic (PLEG): container finished" podID="3c0ee6c9-eb8b-4054-a552-fccd55b68cfc" containerID="f16cb524ca109076a4f65f8865abffeccd6eea5121996937466af337d905c0ae" exitCode=0 Oct 10 16:32:42 crc kubenswrapper[4788]: I1010 16:32:42.887537 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" event={"ID":"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc","Type":"ContainerDied","Data":"f16cb524ca109076a4f65f8865abffeccd6eea5121996937466af337d905c0ae"} Oct 10 16:32:42 crc kubenswrapper[4788]: I1010 16:32:42.887839 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" event={"ID":"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc","Type":"ContainerStarted","Data":"895c00276a579c667ba654b55d426818a768515e1359295c63b2e825f0cc11a2"} Oct 10 16:32:43 crc kubenswrapper[4788]: I1010 16:32:43.900890 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" event={"ID":"3c0ee6c9-eb8b-4054-a552-fccd55b68cfc","Type":"ContainerStarted","Data":"0d45a8cd2e97ab364d9765fc0e7c530ceed4516cf35cd2847158f80b1d858265"} Oct 10 16:32:43 crc kubenswrapper[4788]: I1010 16:32:43.901406 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:43 crc kubenswrapper[4788]: I1010 16:32:43.928987 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" podStartSLOduration=3.928969693 podStartE2EDuration="3.928969693s" podCreationTimestamp="2025-10-10 16:32:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 16:32:43.92204591 +0000 UTC m=+6466.371761458" watchObservedRunningTime="2025-10-10 16:32:43.928969693 +0000 UTC m=+6466.378685241" Oct 10 16:32:46 crc kubenswrapper[4788]: I1010 16:32:46.283597 4788 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9949c5bf-d9nvn" podUID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.148:5353: i/o timeout" Oct 10 16:32:51 crc kubenswrapper[4788]: I1010 16:32:51.236238 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:32:51 crc kubenswrapper[4788]: E1010 16:32:51.237227 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:32:51 crc kubenswrapper[4788]: I1010 16:32:51.310812 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b74f9b95-7wj79" Oct 10 16:32:51 crc kubenswrapper[4788]: I1010 16:32:51.392828 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c94656c95-77nqv"] Oct 10 16:32:51 crc kubenswrapper[4788]: I1010 16:32:51.393097 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" podUID="28d8736e-200c-4765-b78a-4a3913114f72" containerName="dnsmasq-dns" containerID="cri-o://dbca9f587ea5d14ae555bebda043de8949f825999129119d895fcb47c0f9308b" gracePeriod=10 Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.011721 4788 generic.go:334] "Generic (PLEG): container finished" podID="28d8736e-200c-4765-b78a-4a3913114f72" containerID="dbca9f587ea5d14ae555bebda043de8949f825999129119d895fcb47c0f9308b" exitCode=0 Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.012087 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" event={"ID":"28d8736e-200c-4765-b78a-4a3913114f72","Type":"ContainerDied","Data":"dbca9f587ea5d14ae555bebda043de8949f825999129119d895fcb47c0f9308b"} Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.012113 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" event={"ID":"28d8736e-200c-4765-b78a-4a3913114f72","Type":"ContainerDied","Data":"c6256cbb50ecd6904d50b47522d3ab744dd3d9bc18d0f12d3ce9784b80ce01b9"} Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.012125 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6256cbb50ecd6904d50b47522d3ab744dd3d9bc18d0f12d3ce9784b80ce01b9" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.019371 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.108993 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-openstack-cell1\") pod \"28d8736e-200c-4765-b78a-4a3913114f72\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.109193 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4bcv\" (UniqueName: \"kubernetes.io/projected/28d8736e-200c-4765-b78a-4a3913114f72-kube-api-access-l4bcv\") pod \"28d8736e-200c-4765-b78a-4a3913114f72\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.109250 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-sb\") pod \"28d8736e-200c-4765-b78a-4a3913114f72\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.109277 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-dns-svc\") pod \"28d8736e-200c-4765-b78a-4a3913114f72\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.109314 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-nb\") pod \"28d8736e-200c-4765-b78a-4a3913114f72\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.109400 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-config\") pod \"28d8736e-200c-4765-b78a-4a3913114f72\" (UID: \"28d8736e-200c-4765-b78a-4a3913114f72\") " Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.127809 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28d8736e-200c-4765-b78a-4a3913114f72-kube-api-access-l4bcv" (OuterVolumeSpecName: "kube-api-access-l4bcv") pod "28d8736e-200c-4765-b78a-4a3913114f72" (UID: "28d8736e-200c-4765-b78a-4a3913114f72"). InnerVolumeSpecName "kube-api-access-l4bcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.181283 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "28d8736e-200c-4765-b78a-4a3913114f72" (UID: "28d8736e-200c-4765-b78a-4a3913114f72"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.184060 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "28d8736e-200c-4765-b78a-4a3913114f72" (UID: "28d8736e-200c-4765-b78a-4a3913114f72"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.191979 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28d8736e-200c-4765-b78a-4a3913114f72" (UID: "28d8736e-200c-4765-b78a-4a3913114f72"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.195425 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "28d8736e-200c-4765-b78a-4a3913114f72" (UID: "28d8736e-200c-4765-b78a-4a3913114f72"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.212839 4788 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.213063 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4bcv\" (UniqueName: \"kubernetes.io/projected/28d8736e-200c-4765-b78a-4a3913114f72-kube-api-access-l4bcv\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.213126 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.213200 4788 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.213254 4788 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.212997 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-config" (OuterVolumeSpecName: "config") pod "28d8736e-200c-4765-b78a-4a3913114f72" (UID: "28d8736e-200c-4765-b78a-4a3913114f72"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:32:52 crc kubenswrapper[4788]: I1010 16:32:52.315419 4788 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d8736e-200c-4765-b78a-4a3913114f72-config\") on node \"crc\" DevicePath \"\"" Oct 10 16:32:53 crc kubenswrapper[4788]: I1010 16:32:53.023994 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c94656c95-77nqv" Oct 10 16:32:53 crc kubenswrapper[4788]: I1010 16:32:53.054311 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c94656c95-77nqv"] Oct 10 16:32:53 crc kubenswrapper[4788]: I1010 16:32:53.062460 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c94656c95-77nqv"] Oct 10 16:32:54 crc kubenswrapper[4788]: I1010 16:32:54.246104 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28d8736e-200c-4765-b78a-4a3913114f72" path="/var/lib/kubelet/pods/28d8736e-200c-4765-b78a-4a3913114f72/volumes" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.807697 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs"] Oct 10 16:33:02 crc kubenswrapper[4788]: E1010 16:33:02.809546 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerName="dnsmasq-dns" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.809572 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerName="dnsmasq-dns" Oct 10 16:33:02 crc kubenswrapper[4788]: E1010 16:33:02.809600 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerName="init" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.809608 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerName="init" Oct 10 16:33:02 crc kubenswrapper[4788]: E1010 16:33:02.809632 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d8736e-200c-4765-b78a-4a3913114f72" containerName="init" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.809643 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d8736e-200c-4765-b78a-4a3913114f72" containerName="init" Oct 10 16:33:02 crc kubenswrapper[4788]: E1010 16:33:02.809675 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d8736e-200c-4765-b78a-4a3913114f72" containerName="dnsmasq-dns" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.809683 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d8736e-200c-4765-b78a-4a3913114f72" containerName="dnsmasq-dns" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.809974 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c5ab940-e2a4-4024-b4aa-172ee2c6eafc" containerName="dnsmasq-dns" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.809992 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d8736e-200c-4765-b78a-4a3913114f72" containerName="dnsmasq-dns" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.811256 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.813882 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.814661 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.815245 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.815410 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.847525 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs"] Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.903666 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.903749 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.903791 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.904108 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6fc7\" (UniqueName: \"kubernetes.io/projected/17d65674-bacd-4214-a197-2b78b4f9f45d-kube-api-access-k6fc7\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:02 crc kubenswrapper[4788]: I1010 16:33:02.904345 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.006654 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6fc7\" (UniqueName: \"kubernetes.io/projected/17d65674-bacd-4214-a197-2b78b4f9f45d-kube-api-access-k6fc7\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.006740 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.006856 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.006936 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.006977 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.013098 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.013425 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.013548 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.017991 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.029521 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6fc7\" (UniqueName: \"kubernetes.io/projected/17d65674-bacd-4214-a197-2b78b4f9f45d-kube-api-access-k6fc7\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.151372 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:03 crc kubenswrapper[4788]: W1010 16:33:03.782807 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17d65674_bacd_4214_a197_2b78b4f9f45d.slice/crio-9890ffcb1338b66df84e1af0afc968b2855a6d05b845915ebdfb2dcbc8946351 WatchSource:0}: Error finding container 9890ffcb1338b66df84e1af0afc968b2855a6d05b845915ebdfb2dcbc8946351: Status 404 returned error can't find the container with id 9890ffcb1338b66df84e1af0afc968b2855a6d05b845915ebdfb2dcbc8946351 Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.784787 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs"] Oct 10 16:33:03 crc kubenswrapper[4788]: I1010 16:33:03.786195 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:33:04 crc kubenswrapper[4788]: I1010 16:33:04.163358 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" event={"ID":"17d65674-bacd-4214-a197-2b78b4f9f45d","Type":"ContainerStarted","Data":"9890ffcb1338b66df84e1af0afc968b2855a6d05b845915ebdfb2dcbc8946351"} Oct 10 16:33:05 crc kubenswrapper[4788]: I1010 16:33:05.233638 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:33:05 crc kubenswrapper[4788]: E1010 16:33:05.234264 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:33:17 crc kubenswrapper[4788]: I1010 16:33:17.321105 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" event={"ID":"17d65674-bacd-4214-a197-2b78b4f9f45d","Type":"ContainerStarted","Data":"2a3c16d1eb003ae2e3e32cff516aa6133764476947131ef385e1555ddd6dec0b"} Oct 10 16:33:17 crc kubenswrapper[4788]: I1010 16:33:17.341072 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" podStartSLOduration=2.9305679270000002 podStartE2EDuration="15.341050301s" podCreationTimestamp="2025-10-10 16:33:02 +0000 UTC" firstStartedPulling="2025-10-10 16:33:03.785956457 +0000 UTC m=+6486.235672005" lastFinishedPulling="2025-10-10 16:33:16.196438831 +0000 UTC m=+6498.646154379" observedRunningTime="2025-10-10 16:33:17.336987883 +0000 UTC m=+6499.786703441" watchObservedRunningTime="2025-10-10 16:33:17.341050301 +0000 UTC m=+6499.790765849" Oct 10 16:33:18 crc kubenswrapper[4788]: I1010 16:33:18.256568 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:33:18 crc kubenswrapper[4788]: E1010 16:33:18.262711 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:33:29 crc kubenswrapper[4788]: I1010 16:33:29.478060 4788 generic.go:334] "Generic (PLEG): container finished" podID="17d65674-bacd-4214-a197-2b78b4f9f45d" containerID="2a3c16d1eb003ae2e3e32cff516aa6133764476947131ef385e1555ddd6dec0b" exitCode=0 Oct 10 16:33:29 crc kubenswrapper[4788]: I1010 16:33:29.478172 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" event={"ID":"17d65674-bacd-4214-a197-2b78b4f9f45d","Type":"ContainerDied","Data":"2a3c16d1eb003ae2e3e32cff516aa6133764476947131ef385e1555ddd6dec0b"} Oct 10 16:33:30 crc kubenswrapper[4788]: I1010 16:33:30.961381 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.102076 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6fc7\" (UniqueName: \"kubernetes.io/projected/17d65674-bacd-4214-a197-2b78b4f9f45d-kube-api-access-k6fc7\") pod \"17d65674-bacd-4214-a197-2b78b4f9f45d\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.102282 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-inventory\") pod \"17d65674-bacd-4214-a197-2b78b4f9f45d\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.102406 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ceph\") pod \"17d65674-bacd-4214-a197-2b78b4f9f45d\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.102502 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ssh-key\") pod \"17d65674-bacd-4214-a197-2b78b4f9f45d\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.102610 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-pre-adoption-validation-combined-ca-bundle\") pod \"17d65674-bacd-4214-a197-2b78b4f9f45d\" (UID: \"17d65674-bacd-4214-a197-2b78b4f9f45d\") " Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.108742 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "17d65674-bacd-4214-a197-2b78b4f9f45d" (UID: "17d65674-bacd-4214-a197-2b78b4f9f45d"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.109996 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d65674-bacd-4214-a197-2b78b4f9f45d-kube-api-access-k6fc7" (OuterVolumeSpecName: "kube-api-access-k6fc7") pod "17d65674-bacd-4214-a197-2b78b4f9f45d" (UID: "17d65674-bacd-4214-a197-2b78b4f9f45d"). InnerVolumeSpecName "kube-api-access-k6fc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.110185 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ceph" (OuterVolumeSpecName: "ceph") pod "17d65674-bacd-4214-a197-2b78b4f9f45d" (UID: "17d65674-bacd-4214-a197-2b78b4f9f45d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.135371 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-inventory" (OuterVolumeSpecName: "inventory") pod "17d65674-bacd-4214-a197-2b78b4f9f45d" (UID: "17d65674-bacd-4214-a197-2b78b4f9f45d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.137494 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "17d65674-bacd-4214-a197-2b78b4f9f45d" (UID: "17d65674-bacd-4214-a197-2b78b4f9f45d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.205796 4788 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.205839 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6fc7\" (UniqueName: \"kubernetes.io/projected/17d65674-bacd-4214-a197-2b78b4f9f45d-kube-api-access-k6fc7\") on node \"crc\" DevicePath \"\"" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.205855 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.205863 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.205890 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17d65674-bacd-4214-a197-2b78b4f9f45d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.502681 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" event={"ID":"17d65674-bacd-4214-a197-2b78b4f9f45d","Type":"ContainerDied","Data":"9890ffcb1338b66df84e1af0afc968b2855a6d05b845915ebdfb2dcbc8946351"} Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.502723 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs" Oct 10 16:33:31 crc kubenswrapper[4788]: I1010 16:33:31.502737 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9890ffcb1338b66df84e1af0afc968b2855a6d05b845915ebdfb2dcbc8946351" Oct 10 16:33:32 crc kubenswrapper[4788]: I1010 16:33:32.238419 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:33:32 crc kubenswrapper[4788]: E1010 16:33:32.239283 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.328413 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q"] Oct 10 16:33:36 crc kubenswrapper[4788]: E1010 16:33:36.329534 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d65674-bacd-4214-a197-2b78b4f9f45d" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.329550 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d65674-bacd-4214-a197-2b78b4f9f45d" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.329779 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d65674-bacd-4214-a197-2b78b4f9f45d" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.330628 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.333225 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.333311 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.333524 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.334047 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.339885 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q"] Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.422029 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.422154 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nfx7\" (UniqueName: \"kubernetes.io/projected/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-kube-api-access-8nfx7\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.422370 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.422500 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.422578 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.524438 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.524513 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nfx7\" (UniqueName: \"kubernetes.io/projected/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-kube-api-access-8nfx7\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.524541 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.524574 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.524601 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.531883 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.532759 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.533101 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.542178 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.543321 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nfx7\" (UniqueName: \"kubernetes.io/projected/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-kube-api-access-8nfx7\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:36 crc kubenswrapper[4788]: I1010 16:33:36.653379 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.025719 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rxbz9"] Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.029420 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.067996 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rxbz9"] Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.137912 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-utilities\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.138029 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxvd5\" (UniqueName: \"kubernetes.io/projected/21b23cad-2820-4e03-86e1-b5660fffafa6-kube-api-access-jxvd5\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.138362 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-catalog-content\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.211331 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q"] Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.240909 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-utilities\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.240997 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxvd5\" (UniqueName: \"kubernetes.io/projected/21b23cad-2820-4e03-86e1-b5660fffafa6-kube-api-access-jxvd5\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.241062 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-catalog-content\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.241628 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-catalog-content\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.241888 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-utilities\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.264487 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxvd5\" (UniqueName: \"kubernetes.io/projected/21b23cad-2820-4e03-86e1-b5660fffafa6-kube-api-access-jxvd5\") pod \"certified-operators-rxbz9\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.358178 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.569297 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" event={"ID":"6dbab995-d2d3-42cf-8f11-1484c7d00f4c","Type":"ContainerStarted","Data":"b85650119e85693ff750ba0b62e2dc04777d3b1d6d09981281f4a1ad367c5a35"} Oct 10 16:33:37 crc kubenswrapper[4788]: I1010 16:33:37.866149 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rxbz9"] Oct 10 16:33:37 crc kubenswrapper[4788]: W1010 16:33:37.866626 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21b23cad_2820_4e03_86e1_b5660fffafa6.slice/crio-439c559ec4c7964a57522c097680f10bcdcc301f39326abf4ac58568e41e76a9 WatchSource:0}: Error finding container 439c559ec4c7964a57522c097680f10bcdcc301f39326abf4ac58568e41e76a9: Status 404 returned error can't find the container with id 439c559ec4c7964a57522c097680f10bcdcc301f39326abf4ac58568e41e76a9 Oct 10 16:33:38 crc kubenswrapper[4788]: I1010 16:33:38.580315 4788 generic.go:334] "Generic (PLEG): container finished" podID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerID="377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4" exitCode=0 Oct 10 16:33:38 crc kubenswrapper[4788]: I1010 16:33:38.580666 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxbz9" event={"ID":"21b23cad-2820-4e03-86e1-b5660fffafa6","Type":"ContainerDied","Data":"377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4"} Oct 10 16:33:38 crc kubenswrapper[4788]: I1010 16:33:38.580781 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxbz9" event={"ID":"21b23cad-2820-4e03-86e1-b5660fffafa6","Type":"ContainerStarted","Data":"439c559ec4c7964a57522c097680f10bcdcc301f39326abf4ac58568e41e76a9"} Oct 10 16:33:38 crc kubenswrapper[4788]: I1010 16:33:38.582945 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" event={"ID":"6dbab995-d2d3-42cf-8f11-1484c7d00f4c","Type":"ContainerStarted","Data":"444d335fffb6499fdeefd39e412e1c7533c22be9568bc84f96fe066bba5923c5"} Oct 10 16:33:38 crc kubenswrapper[4788]: I1010 16:33:38.627713 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" podStartSLOduration=2.407461072 podStartE2EDuration="2.627689467s" podCreationTimestamp="2025-10-10 16:33:36 +0000 UTC" firstStartedPulling="2025-10-10 16:33:37.208715014 +0000 UTC m=+6519.658430562" lastFinishedPulling="2025-10-10 16:33:37.428943409 +0000 UTC m=+6519.878658957" observedRunningTime="2025-10-10 16:33:38.622481529 +0000 UTC m=+6521.072197077" watchObservedRunningTime="2025-10-10 16:33:38.627689467 +0000 UTC m=+6521.077405035" Oct 10 16:33:39 crc kubenswrapper[4788]: I1010 16:33:39.043800 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-hnbp6"] Oct 10 16:33:39 crc kubenswrapper[4788]: I1010 16:33:39.059572 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-hnbp6"] Oct 10 16:33:39 crc kubenswrapper[4788]: I1010 16:33:39.596834 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxbz9" event={"ID":"21b23cad-2820-4e03-86e1-b5660fffafa6","Type":"ContainerStarted","Data":"c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449"} Oct 10 16:33:40 crc kubenswrapper[4788]: I1010 16:33:40.246805 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab" path="/var/lib/kubelet/pods/36dc4aea-cb0e-4767-b8bf-0d9ce4ffc9ab/volumes" Oct 10 16:33:41 crc kubenswrapper[4788]: I1010 16:33:41.637109 4788 generic.go:334] "Generic (PLEG): container finished" podID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerID="c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449" exitCode=0 Oct 10 16:33:41 crc kubenswrapper[4788]: I1010 16:33:41.637240 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxbz9" event={"ID":"21b23cad-2820-4e03-86e1-b5660fffafa6","Type":"ContainerDied","Data":"c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449"} Oct 10 16:33:42 crc kubenswrapper[4788]: I1010 16:33:42.650310 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxbz9" event={"ID":"21b23cad-2820-4e03-86e1-b5660fffafa6","Type":"ContainerStarted","Data":"7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9"} Oct 10 16:33:42 crc kubenswrapper[4788]: I1010 16:33:42.688259 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rxbz9" podStartSLOduration=3.027583695 podStartE2EDuration="6.688242711s" podCreationTimestamp="2025-10-10 16:33:36 +0000 UTC" firstStartedPulling="2025-10-10 16:33:38.584127537 +0000 UTC m=+6521.033843085" lastFinishedPulling="2025-10-10 16:33:42.244786543 +0000 UTC m=+6524.694502101" observedRunningTime="2025-10-10 16:33:42.684685937 +0000 UTC m=+6525.134401515" watchObservedRunningTime="2025-10-10 16:33:42.688242711 +0000 UTC m=+6525.137958259" Oct 10 16:33:47 crc kubenswrapper[4788]: I1010 16:33:47.235340 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:33:47 crc kubenswrapper[4788]: E1010 16:33:47.236414 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:33:47 crc kubenswrapper[4788]: I1010 16:33:47.358494 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:47 crc kubenswrapper[4788]: I1010 16:33:47.359037 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:47 crc kubenswrapper[4788]: I1010 16:33:47.420814 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:47 crc kubenswrapper[4788]: I1010 16:33:47.759045 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:47 crc kubenswrapper[4788]: I1010 16:33:47.831212 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rxbz9"] Oct 10 16:33:49 crc kubenswrapper[4788]: I1010 16:33:49.737416 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rxbz9" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerName="registry-server" containerID="cri-o://7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9" gracePeriod=2 Oct 10 16:33:49 crc kubenswrapper[4788]: E1010 16:33:49.906128 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21b23cad_2820_4e03_86e1_b5660fffafa6.slice/crio-7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9.scope\": RecentStats: unable to find data in memory cache]" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.316949 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.497190 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-utilities\") pod \"21b23cad-2820-4e03-86e1-b5660fffafa6\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.497250 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxvd5\" (UniqueName: \"kubernetes.io/projected/21b23cad-2820-4e03-86e1-b5660fffafa6-kube-api-access-jxvd5\") pod \"21b23cad-2820-4e03-86e1-b5660fffafa6\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.497313 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-catalog-content\") pod \"21b23cad-2820-4e03-86e1-b5660fffafa6\" (UID: \"21b23cad-2820-4e03-86e1-b5660fffafa6\") " Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.498859 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-utilities" (OuterVolumeSpecName: "utilities") pod "21b23cad-2820-4e03-86e1-b5660fffafa6" (UID: "21b23cad-2820-4e03-86e1-b5660fffafa6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.506385 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21b23cad-2820-4e03-86e1-b5660fffafa6-kube-api-access-jxvd5" (OuterVolumeSpecName: "kube-api-access-jxvd5") pod "21b23cad-2820-4e03-86e1-b5660fffafa6" (UID: "21b23cad-2820-4e03-86e1-b5660fffafa6"). InnerVolumeSpecName "kube-api-access-jxvd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.555630 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21b23cad-2820-4e03-86e1-b5660fffafa6" (UID: "21b23cad-2820-4e03-86e1-b5660fffafa6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.600258 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.600303 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxvd5\" (UniqueName: \"kubernetes.io/projected/21b23cad-2820-4e03-86e1-b5660fffafa6-kube-api-access-jxvd5\") on node \"crc\" DevicePath \"\"" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.600317 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21b23cad-2820-4e03-86e1-b5660fffafa6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.751354 4788 generic.go:334] "Generic (PLEG): container finished" podID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerID="7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9" exitCode=0 Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.751410 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxbz9" event={"ID":"21b23cad-2820-4e03-86e1-b5660fffafa6","Type":"ContainerDied","Data":"7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9"} Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.751460 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rxbz9" event={"ID":"21b23cad-2820-4e03-86e1-b5660fffafa6","Type":"ContainerDied","Data":"439c559ec4c7964a57522c097680f10bcdcc301f39326abf4ac58568e41e76a9"} Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.751483 4788 scope.go:117] "RemoveContainer" containerID="7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.752500 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rxbz9" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.807462 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rxbz9"] Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.808294 4788 scope.go:117] "RemoveContainer" containerID="c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.820278 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rxbz9"] Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.830531 4788 scope.go:117] "RemoveContainer" containerID="377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.878381 4788 scope.go:117] "RemoveContainer" containerID="7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9" Oct 10 16:33:50 crc kubenswrapper[4788]: E1010 16:33:50.878915 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9\": container with ID starting with 7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9 not found: ID does not exist" containerID="7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.878954 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9"} err="failed to get container status \"7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9\": rpc error: code = NotFound desc = could not find container \"7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9\": container with ID starting with 7dd9092128d110edd1ab88a7caca264e788faf56cc3b89a38705b8ad7a5ca2f9 not found: ID does not exist" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.878980 4788 scope.go:117] "RemoveContainer" containerID="c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449" Oct 10 16:33:50 crc kubenswrapper[4788]: E1010 16:33:50.879662 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449\": container with ID starting with c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449 not found: ID does not exist" containerID="c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.879748 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449"} err="failed to get container status \"c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449\": rpc error: code = NotFound desc = could not find container \"c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449\": container with ID starting with c45b424e6b0c702bb42fc0b2ea1cdee625ef1ad1111fc921e3124e7030ab2449 not found: ID does not exist" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.879803 4788 scope.go:117] "RemoveContainer" containerID="377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4" Oct 10 16:33:50 crc kubenswrapper[4788]: E1010 16:33:50.880332 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4\": container with ID starting with 377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4 not found: ID does not exist" containerID="377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4" Oct 10 16:33:50 crc kubenswrapper[4788]: I1010 16:33:50.880384 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4"} err="failed to get container status \"377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4\": rpc error: code = NotFound desc = could not find container \"377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4\": container with ID starting with 377bc09a305600d0c6315fc9006345adcb88ce77c8e6f74d0f0b9aece8ed18a4 not found: ID does not exist" Oct 10 16:33:51 crc kubenswrapper[4788]: I1010 16:33:51.056202 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-725d-account-create-h76lw"] Oct 10 16:33:51 crc kubenswrapper[4788]: I1010 16:33:51.066956 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-725d-account-create-h76lw"] Oct 10 16:33:52 crc kubenswrapper[4788]: I1010 16:33:52.251152 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" path="/var/lib/kubelet/pods/21b23cad-2820-4e03-86e1-b5660fffafa6/volumes" Oct 10 16:33:52 crc kubenswrapper[4788]: I1010 16:33:52.252383 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b437783d-adf5-442e-ba13-3f2abd872426" path="/var/lib/kubelet/pods/b437783d-adf5-442e-ba13-3f2abd872426/volumes" Oct 10 16:33:59 crc kubenswrapper[4788]: I1010 16:33:59.036248 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-qwtnc"] Oct 10 16:33:59 crc kubenswrapper[4788]: I1010 16:33:59.046450 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-qwtnc"] Oct 10 16:34:00 crc kubenswrapper[4788]: I1010 16:34:00.260713 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25598815-a0ed-4d38-9e06-89bac3efd642" path="/var/lib/kubelet/pods/25598815-a0ed-4d38-9e06-89bac3efd642/volumes" Oct 10 16:34:01 crc kubenswrapper[4788]: I1010 16:34:01.235505 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:34:01 crc kubenswrapper[4788]: E1010 16:34:01.236060 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:34:10 crc kubenswrapper[4788]: I1010 16:34:10.046293 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-1c0e-account-create-lmchj"] Oct 10 16:34:10 crc kubenswrapper[4788]: I1010 16:34:10.060535 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-1c0e-account-create-lmchj"] Oct 10 16:34:10 crc kubenswrapper[4788]: I1010 16:34:10.248955 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd8c72b5-9106-40ec-9b4e-6d993a0799c5" path="/var/lib/kubelet/pods/fd8c72b5-9106-40ec-9b4e-6d993a0799c5/volumes" Oct 10 16:34:12 crc kubenswrapper[4788]: I1010 16:34:12.235156 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:34:12 crc kubenswrapper[4788]: E1010 16:34:12.235781 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:34:18 crc kubenswrapper[4788]: I1010 16:34:18.244033 4788 scope.go:117] "RemoveContainer" containerID="7efe00d56a6fbeabf4a56a72fe9e1377374e2723ed9c354311e0d50558d5a5de" Oct 10 16:34:18 crc kubenswrapper[4788]: I1010 16:34:18.279360 4788 scope.go:117] "RemoveContainer" containerID="5b5bd771151fef05295f13aef1233a50d12abbc8a7b111e015e9bfaf563aa042" Oct 10 16:34:18 crc kubenswrapper[4788]: I1010 16:34:18.336648 4788 scope.go:117] "RemoveContainer" containerID="39683b019f3f5c7bc838991615f91a9f353885734145b9ed3cec5c378499a95c" Oct 10 16:34:18 crc kubenswrapper[4788]: I1010 16:34:18.400993 4788 scope.go:117] "RemoveContainer" containerID="9f166804d7756e8bfe9c780cc68371874fab9f8a5c1ac77e3d595d81fd1b45ee" Oct 10 16:34:23 crc kubenswrapper[4788]: I1010 16:34:23.242906 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:34:23 crc kubenswrapper[4788]: E1010 16:34:23.243654 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:34:25 crc kubenswrapper[4788]: I1010 16:34:25.938932 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xlwsx"] Oct 10 16:34:25 crc kubenswrapper[4788]: E1010 16:34:25.940497 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerName="extract-utilities" Oct 10 16:34:25 crc kubenswrapper[4788]: I1010 16:34:25.940517 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerName="extract-utilities" Oct 10 16:34:25 crc kubenswrapper[4788]: E1010 16:34:25.940547 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerName="extract-content" Oct 10 16:34:25 crc kubenswrapper[4788]: I1010 16:34:25.940554 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerName="extract-content" Oct 10 16:34:25 crc kubenswrapper[4788]: E1010 16:34:25.940606 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerName="registry-server" Oct 10 16:34:25 crc kubenswrapper[4788]: I1010 16:34:25.940616 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerName="registry-server" Oct 10 16:34:25 crc kubenswrapper[4788]: I1010 16:34:25.940944 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="21b23cad-2820-4e03-86e1-b5660fffafa6" containerName="registry-server" Oct 10 16:34:25 crc kubenswrapper[4788]: I1010 16:34:25.943468 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:25 crc kubenswrapper[4788]: I1010 16:34:25.964593 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlwsx"] Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.045417 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-catalog-content\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.045612 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmfzv\" (UniqueName: \"kubernetes.io/projected/95272bfd-5785-42a9-8d58-b0e35aa787e3-kube-api-access-tmfzv\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.045786 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-utilities\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.148331 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-catalog-content\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.148448 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmfzv\" (UniqueName: \"kubernetes.io/projected/95272bfd-5785-42a9-8d58-b0e35aa787e3-kube-api-access-tmfzv\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.148537 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-utilities\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.149121 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-utilities\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.149363 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-catalog-content\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.168204 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmfzv\" (UniqueName: \"kubernetes.io/projected/95272bfd-5785-42a9-8d58-b0e35aa787e3-kube-api-access-tmfzv\") pod \"redhat-marketplace-xlwsx\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.270703 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:26 crc kubenswrapper[4788]: I1010 16:34:26.802186 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlwsx"] Oct 10 16:34:27 crc kubenswrapper[4788]: I1010 16:34:27.199962 4788 generic.go:334] "Generic (PLEG): container finished" podID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerID="fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b" exitCode=0 Oct 10 16:34:27 crc kubenswrapper[4788]: I1010 16:34:27.200035 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlwsx" event={"ID":"95272bfd-5785-42a9-8d58-b0e35aa787e3","Type":"ContainerDied","Data":"fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b"} Oct 10 16:34:27 crc kubenswrapper[4788]: I1010 16:34:27.200305 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlwsx" event={"ID":"95272bfd-5785-42a9-8d58-b0e35aa787e3","Type":"ContainerStarted","Data":"d91e9c9282e9c84dfa52e49a1dfac7ac119eb5ade0e69ea50448dc1d6a401f1f"} Oct 10 16:34:28 crc kubenswrapper[4788]: I1010 16:34:28.211018 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlwsx" event={"ID":"95272bfd-5785-42a9-8d58-b0e35aa787e3","Type":"ContainerStarted","Data":"a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8"} Oct 10 16:34:29 crc kubenswrapper[4788]: I1010 16:34:29.224395 4788 generic.go:334] "Generic (PLEG): container finished" podID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerID="a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8" exitCode=0 Oct 10 16:34:29 crc kubenswrapper[4788]: I1010 16:34:29.224443 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlwsx" event={"ID":"95272bfd-5785-42a9-8d58-b0e35aa787e3","Type":"ContainerDied","Data":"a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8"} Oct 10 16:34:30 crc kubenswrapper[4788]: I1010 16:34:30.254206 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlwsx" event={"ID":"95272bfd-5785-42a9-8d58-b0e35aa787e3","Type":"ContainerStarted","Data":"8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e"} Oct 10 16:34:30 crc kubenswrapper[4788]: I1010 16:34:30.278271 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xlwsx" podStartSLOduration=2.816514563 podStartE2EDuration="5.278234925s" podCreationTimestamp="2025-10-10 16:34:25 +0000 UTC" firstStartedPulling="2025-10-10 16:34:27.203022706 +0000 UTC m=+6569.652738274" lastFinishedPulling="2025-10-10 16:34:29.664743078 +0000 UTC m=+6572.114458636" observedRunningTime="2025-10-10 16:34:30.270923602 +0000 UTC m=+6572.720639150" watchObservedRunningTime="2025-10-10 16:34:30.278234925 +0000 UTC m=+6572.727950473" Oct 10 16:34:36 crc kubenswrapper[4788]: I1010 16:34:36.271898 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:36 crc kubenswrapper[4788]: I1010 16:34:36.273498 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:36 crc kubenswrapper[4788]: I1010 16:34:36.327369 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:36 crc kubenswrapper[4788]: I1010 16:34:36.401008 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:36 crc kubenswrapper[4788]: I1010 16:34:36.615186 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlwsx"] Oct 10 16:34:38 crc kubenswrapper[4788]: I1010 16:34:38.250595 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:34:38 crc kubenswrapper[4788]: E1010 16:34:38.250915 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:34:38 crc kubenswrapper[4788]: I1010 16:34:38.370941 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xlwsx" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerName="registry-server" containerID="cri-o://8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e" gracePeriod=2 Oct 10 16:34:38 crc kubenswrapper[4788]: I1010 16:34:38.887374 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:38 crc kubenswrapper[4788]: I1010 16:34:38.987193 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-utilities\") pod \"95272bfd-5785-42a9-8d58-b0e35aa787e3\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " Oct 10 16:34:38 crc kubenswrapper[4788]: I1010 16:34:38.987339 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmfzv\" (UniqueName: \"kubernetes.io/projected/95272bfd-5785-42a9-8d58-b0e35aa787e3-kube-api-access-tmfzv\") pod \"95272bfd-5785-42a9-8d58-b0e35aa787e3\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " Oct 10 16:34:38 crc kubenswrapper[4788]: I1010 16:34:38.987437 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-catalog-content\") pod \"95272bfd-5785-42a9-8d58-b0e35aa787e3\" (UID: \"95272bfd-5785-42a9-8d58-b0e35aa787e3\") " Oct 10 16:34:38 crc kubenswrapper[4788]: I1010 16:34:38.988134 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-utilities" (OuterVolumeSpecName: "utilities") pod "95272bfd-5785-42a9-8d58-b0e35aa787e3" (UID: "95272bfd-5785-42a9-8d58-b0e35aa787e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:34:38 crc kubenswrapper[4788]: I1010 16:34:38.994715 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95272bfd-5785-42a9-8d58-b0e35aa787e3-kube-api-access-tmfzv" (OuterVolumeSpecName: "kube-api-access-tmfzv") pod "95272bfd-5785-42a9-8d58-b0e35aa787e3" (UID: "95272bfd-5785-42a9-8d58-b0e35aa787e3"). InnerVolumeSpecName "kube-api-access-tmfzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.005394 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95272bfd-5785-42a9-8d58-b0e35aa787e3" (UID: "95272bfd-5785-42a9-8d58-b0e35aa787e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.090374 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.090412 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmfzv\" (UniqueName: \"kubernetes.io/projected/95272bfd-5785-42a9-8d58-b0e35aa787e3-kube-api-access-tmfzv\") on node \"crc\" DevicePath \"\"" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.090422 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95272bfd-5785-42a9-8d58-b0e35aa787e3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.383357 4788 generic.go:334] "Generic (PLEG): container finished" podID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerID="8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e" exitCode=0 Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.383465 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlwsx" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.383469 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlwsx" event={"ID":"95272bfd-5785-42a9-8d58-b0e35aa787e3","Type":"ContainerDied","Data":"8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e"} Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.383841 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlwsx" event={"ID":"95272bfd-5785-42a9-8d58-b0e35aa787e3","Type":"ContainerDied","Data":"d91e9c9282e9c84dfa52e49a1dfac7ac119eb5ade0e69ea50448dc1d6a401f1f"} Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.383866 4788 scope.go:117] "RemoveContainer" containerID="8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.409378 4788 scope.go:117] "RemoveContainer" containerID="a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.436404 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlwsx"] Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.438824 4788 scope.go:117] "RemoveContainer" containerID="fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.449100 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlwsx"] Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.499515 4788 scope.go:117] "RemoveContainer" containerID="8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e" Oct 10 16:34:39 crc kubenswrapper[4788]: E1010 16:34:39.499982 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e\": container with ID starting with 8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e not found: ID does not exist" containerID="8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.500035 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e"} err="failed to get container status \"8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e\": rpc error: code = NotFound desc = could not find container \"8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e\": container with ID starting with 8dc5a530fc0f336d5d6a45d9e4927e1a0e82c24dd961396bc86bcd35b127555e not found: ID does not exist" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.500065 4788 scope.go:117] "RemoveContainer" containerID="a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8" Oct 10 16:34:39 crc kubenswrapper[4788]: E1010 16:34:39.500589 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8\": container with ID starting with a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8 not found: ID does not exist" containerID="a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.500701 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8"} err="failed to get container status \"a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8\": rpc error: code = NotFound desc = could not find container \"a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8\": container with ID starting with a374c53012ba76f0cda0eebd66937bd31e38711e819daaffd5862296831e67c8 not found: ID does not exist" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.500751 4788 scope.go:117] "RemoveContainer" containerID="fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b" Oct 10 16:34:39 crc kubenswrapper[4788]: E1010 16:34:39.501056 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b\": container with ID starting with fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b not found: ID does not exist" containerID="fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b" Oct 10 16:34:39 crc kubenswrapper[4788]: I1010 16:34:39.501082 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b"} err="failed to get container status \"fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b\": rpc error: code = NotFound desc = could not find container \"fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b\": container with ID starting with fba6a8f3aca695e6cb744c3f4a23da2551d629b5a5bb887364178bd622ade81b not found: ID does not exist" Oct 10 16:34:40 crc kubenswrapper[4788]: I1010 16:34:40.252167 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" path="/var/lib/kubelet/pods/95272bfd-5785-42a9-8d58-b0e35aa787e3/volumes" Oct 10 16:34:52 crc kubenswrapper[4788]: I1010 16:34:52.235098 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:34:52 crc kubenswrapper[4788]: E1010 16:34:52.235876 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:34:54 crc kubenswrapper[4788]: I1010 16:34:54.052793 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-582jp"] Oct 10 16:34:54 crc kubenswrapper[4788]: I1010 16:34:54.066637 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-582jp"] Oct 10 16:34:54 crc kubenswrapper[4788]: I1010 16:34:54.250788 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18752d95-0174-4746-8554-00be4abbcaa7" path="/var/lib/kubelet/pods/18752d95-0174-4746-8554-00be4abbcaa7/volumes" Oct 10 16:35:06 crc kubenswrapper[4788]: I1010 16:35:06.234593 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:35:06 crc kubenswrapper[4788]: I1010 16:35:06.675831 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"55fb8770913a06b4815ab6ba7c27a4726c032dd815a83476336b3d0f44277d9b"} Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.176810 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dwl9r"] Oct 10 16:35:12 crc kubenswrapper[4788]: E1010 16:35:12.195366 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerName="registry-server" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.195387 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerName="registry-server" Oct 10 16:35:12 crc kubenswrapper[4788]: E1010 16:35:12.195432 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerName="extract-utilities" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.195441 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerName="extract-utilities" Oct 10 16:35:12 crc kubenswrapper[4788]: E1010 16:35:12.195451 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerName="extract-content" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.195460 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerName="extract-content" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.195816 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="95272bfd-5785-42a9-8d58-b0e35aa787e3" containerName="registry-server" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.198056 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dwl9r"] Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.198176 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.337790 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-utilities\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.338630 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-catalog-content\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.338960 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpqh4\" (UniqueName: \"kubernetes.io/projected/0ff9d646-585c-44f9-9ac6-86d943b18a27-kube-api-access-vpqh4\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.442533 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-catalog-content\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.442681 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpqh4\" (UniqueName: \"kubernetes.io/projected/0ff9d646-585c-44f9-9ac6-86d943b18a27-kube-api-access-vpqh4\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.442828 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-utilities\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.443240 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-catalog-content\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.443301 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-utilities\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.462209 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpqh4\" (UniqueName: \"kubernetes.io/projected/0ff9d646-585c-44f9-9ac6-86d943b18a27-kube-api-access-vpqh4\") pod \"redhat-operators-dwl9r\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:12 crc kubenswrapper[4788]: I1010 16:35:12.539080 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:13 crc kubenswrapper[4788]: I1010 16:35:13.070729 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dwl9r"] Oct 10 16:35:13 crc kubenswrapper[4788]: I1010 16:35:13.757085 4788 generic.go:334] "Generic (PLEG): container finished" podID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerID="50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82" exitCode=0 Oct 10 16:35:13 crc kubenswrapper[4788]: I1010 16:35:13.757232 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwl9r" event={"ID":"0ff9d646-585c-44f9-9ac6-86d943b18a27","Type":"ContainerDied","Data":"50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82"} Oct 10 16:35:13 crc kubenswrapper[4788]: I1010 16:35:13.757631 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwl9r" event={"ID":"0ff9d646-585c-44f9-9ac6-86d943b18a27","Type":"ContainerStarted","Data":"78a3e5018f8df060859b577d4f568b0c1be0e6071e1a945aecab3e9dcf3cbe0f"} Oct 10 16:35:14 crc kubenswrapper[4788]: I1010 16:35:14.772572 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwl9r" event={"ID":"0ff9d646-585c-44f9-9ac6-86d943b18a27","Type":"ContainerStarted","Data":"1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae"} Oct 10 16:35:17 crc kubenswrapper[4788]: I1010 16:35:17.805359 4788 generic.go:334] "Generic (PLEG): container finished" podID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerID="1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae" exitCode=0 Oct 10 16:35:17 crc kubenswrapper[4788]: I1010 16:35:17.805474 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwl9r" event={"ID":"0ff9d646-585c-44f9-9ac6-86d943b18a27","Type":"ContainerDied","Data":"1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae"} Oct 10 16:35:18 crc kubenswrapper[4788]: I1010 16:35:18.569982 4788 scope.go:117] "RemoveContainer" containerID="d05d0681a13d9545bb1158106066a0fae930cb485a42a6d73ce005e91ee8cee7" Oct 10 16:35:18 crc kubenswrapper[4788]: I1010 16:35:18.665347 4788 scope.go:117] "RemoveContainer" containerID="dec256bd3c52330cd0cce3600810ca50b9258497051c772792dc53bff210c1d9" Oct 10 16:35:18 crc kubenswrapper[4788]: I1010 16:35:18.823277 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwl9r" event={"ID":"0ff9d646-585c-44f9-9ac6-86d943b18a27","Type":"ContainerStarted","Data":"5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa"} Oct 10 16:35:18 crc kubenswrapper[4788]: I1010 16:35:18.856567 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dwl9r" podStartSLOduration=2.389764353 podStartE2EDuration="6.856519313s" podCreationTimestamp="2025-10-10 16:35:12 +0000 UTC" firstStartedPulling="2025-10-10 16:35:13.760993353 +0000 UTC m=+6616.210708901" lastFinishedPulling="2025-10-10 16:35:18.227748313 +0000 UTC m=+6620.677463861" observedRunningTime="2025-10-10 16:35:18.844688941 +0000 UTC m=+6621.294404489" watchObservedRunningTime="2025-10-10 16:35:18.856519313 +0000 UTC m=+6621.306234861" Oct 10 16:35:22 crc kubenswrapper[4788]: I1010 16:35:22.539272 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:22 crc kubenswrapper[4788]: I1010 16:35:22.539908 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:23 crc kubenswrapper[4788]: I1010 16:35:23.596167 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dwl9r" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="registry-server" probeResult="failure" output=< Oct 10 16:35:23 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 16:35:23 crc kubenswrapper[4788]: > Oct 10 16:35:33 crc kubenswrapper[4788]: I1010 16:35:33.603306 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dwl9r" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="registry-server" probeResult="failure" output=< Oct 10 16:35:33 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 16:35:33 crc kubenswrapper[4788]: > Oct 10 16:35:43 crc kubenswrapper[4788]: I1010 16:35:43.635349 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dwl9r" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="registry-server" probeResult="failure" output=< Oct 10 16:35:43 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 16:35:43 crc kubenswrapper[4788]: > Oct 10 16:35:52 crc kubenswrapper[4788]: I1010 16:35:52.600024 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:52 crc kubenswrapper[4788]: I1010 16:35:52.663436 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:53 crc kubenswrapper[4788]: I1010 16:35:53.803863 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dwl9r"] Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.249791 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dwl9r" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="registry-server" containerID="cri-o://5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa" gracePeriod=2 Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.713756 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.808997 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpqh4\" (UniqueName: \"kubernetes.io/projected/0ff9d646-585c-44f9-9ac6-86d943b18a27-kube-api-access-vpqh4\") pod \"0ff9d646-585c-44f9-9ac6-86d943b18a27\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.809053 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-catalog-content\") pod \"0ff9d646-585c-44f9-9ac6-86d943b18a27\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.809225 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-utilities\") pod \"0ff9d646-585c-44f9-9ac6-86d943b18a27\" (UID: \"0ff9d646-585c-44f9-9ac6-86d943b18a27\") " Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.809793 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-utilities" (OuterVolumeSpecName: "utilities") pod "0ff9d646-585c-44f9-9ac6-86d943b18a27" (UID: "0ff9d646-585c-44f9-9ac6-86d943b18a27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.814101 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ff9d646-585c-44f9-9ac6-86d943b18a27-kube-api-access-vpqh4" (OuterVolumeSpecName: "kube-api-access-vpqh4") pod "0ff9d646-585c-44f9-9ac6-86d943b18a27" (UID: "0ff9d646-585c-44f9-9ac6-86d943b18a27"). InnerVolumeSpecName "kube-api-access-vpqh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.911977 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.912039 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpqh4\" (UniqueName: \"kubernetes.io/projected/0ff9d646-585c-44f9-9ac6-86d943b18a27-kube-api-access-vpqh4\") on node \"crc\" DevicePath \"\"" Oct 10 16:35:54 crc kubenswrapper[4788]: I1010 16:35:54.917753 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ff9d646-585c-44f9-9ac6-86d943b18a27" (UID: "0ff9d646-585c-44f9-9ac6-86d943b18a27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.014614 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ff9d646-585c-44f9-9ac6-86d943b18a27-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.265706 4788 generic.go:334] "Generic (PLEG): container finished" podID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerID="5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa" exitCode=0 Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.265744 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwl9r" event={"ID":"0ff9d646-585c-44f9-9ac6-86d943b18a27","Type":"ContainerDied","Data":"5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa"} Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.265770 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dwl9r" event={"ID":"0ff9d646-585c-44f9-9ac6-86d943b18a27","Type":"ContainerDied","Data":"78a3e5018f8df060859b577d4f568b0c1be0e6071e1a945aecab3e9dcf3cbe0f"} Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.265793 4788 scope.go:117] "RemoveContainer" containerID="5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.265890 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dwl9r" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.311220 4788 scope.go:117] "RemoveContainer" containerID="1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.322889 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dwl9r"] Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.344805 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dwl9r"] Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.347432 4788 scope.go:117] "RemoveContainer" containerID="50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.403776 4788 scope.go:117] "RemoveContainer" containerID="5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa" Oct 10 16:35:55 crc kubenswrapper[4788]: E1010 16:35:55.404379 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa\": container with ID starting with 5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa not found: ID does not exist" containerID="5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.404426 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa"} err="failed to get container status \"5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa\": rpc error: code = NotFound desc = could not find container \"5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa\": container with ID starting with 5cf0fb0bb2cb4a5d6d13ada2f25c1757c05a132c5ca615b16c65f5d56b1d07aa not found: ID does not exist" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.404454 4788 scope.go:117] "RemoveContainer" containerID="1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae" Oct 10 16:35:55 crc kubenswrapper[4788]: E1010 16:35:55.404839 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae\": container with ID starting with 1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae not found: ID does not exist" containerID="1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.404866 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae"} err="failed to get container status \"1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae\": rpc error: code = NotFound desc = could not find container \"1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae\": container with ID starting with 1aa36fcc2d7398dfd7e759dbd898359c523c8beba1c06d2caa10e5030aefdfae not found: ID does not exist" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.404882 4788 scope.go:117] "RemoveContainer" containerID="50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82" Oct 10 16:35:55 crc kubenswrapper[4788]: E1010 16:35:55.405178 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82\": container with ID starting with 50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82 not found: ID does not exist" containerID="50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82" Oct 10 16:35:55 crc kubenswrapper[4788]: I1010 16:35:55.405203 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82"} err="failed to get container status \"50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82\": rpc error: code = NotFound desc = could not find container \"50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82\": container with ID starting with 50d6d77c67254bea32669372071a2c4c9260fb12471cb0a44627a0019ab07e82 not found: ID does not exist" Oct 10 16:35:56 crc kubenswrapper[4788]: I1010 16:35:56.252934 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" path="/var/lib/kubelet/pods/0ff9d646-585c-44f9-9ac6-86d943b18a27/volumes" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.661064 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4zllk"] Oct 10 16:36:56 crc kubenswrapper[4788]: E1010 16:36:56.662072 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="extract-utilities" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.662159 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="extract-utilities" Oct 10 16:36:56 crc kubenswrapper[4788]: E1010 16:36:56.662206 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="registry-server" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.662212 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="registry-server" Oct 10 16:36:56 crc kubenswrapper[4788]: E1010 16:36:56.662220 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="extract-content" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.662226 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="extract-content" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.662431 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ff9d646-585c-44f9-9ac6-86d943b18a27" containerName="registry-server" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.664037 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.683730 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4zllk"] Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.758420 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmp29\" (UniqueName: \"kubernetes.io/projected/9a06cbe8-94ad-4a13-b29a-500768bb42fa-kube-api-access-pmp29\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.758518 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-utilities\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.758545 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-catalog-content\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.862331 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmp29\" (UniqueName: \"kubernetes.io/projected/9a06cbe8-94ad-4a13-b29a-500768bb42fa-kube-api-access-pmp29\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.862676 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-utilities\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.862785 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-catalog-content\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.863398 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-utilities\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.863463 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-catalog-content\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.886298 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmp29\" (UniqueName: \"kubernetes.io/projected/9a06cbe8-94ad-4a13-b29a-500768bb42fa-kube-api-access-pmp29\") pod \"community-operators-4zllk\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:56 crc kubenswrapper[4788]: I1010 16:36:56.987807 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:36:57 crc kubenswrapper[4788]: I1010 16:36:57.563500 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4zllk"] Oct 10 16:36:58 crc kubenswrapper[4788]: I1010 16:36:58.013638 4788 generic.go:334] "Generic (PLEG): container finished" podID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerID="95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f" exitCode=0 Oct 10 16:36:58 crc kubenswrapper[4788]: I1010 16:36:58.013760 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4zllk" event={"ID":"9a06cbe8-94ad-4a13-b29a-500768bb42fa","Type":"ContainerDied","Data":"95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f"} Oct 10 16:36:58 crc kubenswrapper[4788]: I1010 16:36:58.014137 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4zllk" event={"ID":"9a06cbe8-94ad-4a13-b29a-500768bb42fa","Type":"ContainerStarted","Data":"df3b29d71298d4f75a7a57350d8e342976d65065e340f3681c01afcc3fee57f4"} Oct 10 16:36:59 crc kubenswrapper[4788]: I1010 16:36:59.027638 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4zllk" event={"ID":"9a06cbe8-94ad-4a13-b29a-500768bb42fa","Type":"ContainerStarted","Data":"fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b"} Oct 10 16:37:01 crc kubenswrapper[4788]: I1010 16:37:01.059714 4788 generic.go:334] "Generic (PLEG): container finished" podID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerID="fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b" exitCode=0 Oct 10 16:37:01 crc kubenswrapper[4788]: I1010 16:37:01.059783 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4zllk" event={"ID":"9a06cbe8-94ad-4a13-b29a-500768bb42fa","Type":"ContainerDied","Data":"fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b"} Oct 10 16:37:02 crc kubenswrapper[4788]: I1010 16:37:02.081737 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4zllk" event={"ID":"9a06cbe8-94ad-4a13-b29a-500768bb42fa","Type":"ContainerStarted","Data":"23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e"} Oct 10 16:37:02 crc kubenswrapper[4788]: I1010 16:37:02.106396 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4zllk" podStartSLOduration=2.597136855 podStartE2EDuration="6.106373624s" podCreationTimestamp="2025-10-10 16:36:56 +0000 UTC" firstStartedPulling="2025-10-10 16:36:58.015837088 +0000 UTC m=+6720.465552646" lastFinishedPulling="2025-10-10 16:37:01.525073837 +0000 UTC m=+6723.974789415" observedRunningTime="2025-10-10 16:37:02.102707497 +0000 UTC m=+6724.552423045" watchObservedRunningTime="2025-10-10 16:37:02.106373624 +0000 UTC m=+6724.556089172" Oct 10 16:37:06 crc kubenswrapper[4788]: I1010 16:37:06.988229 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:37:06 crc kubenswrapper[4788]: I1010 16:37:06.989130 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:37:07 crc kubenswrapper[4788]: I1010 16:37:07.072233 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:37:07 crc kubenswrapper[4788]: I1010 16:37:07.207477 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:37:07 crc kubenswrapper[4788]: I1010 16:37:07.333706 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4zllk"] Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.165742 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4zllk" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerName="registry-server" containerID="cri-o://23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e" gracePeriod=2 Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.785882 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.879873 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmp29\" (UniqueName: \"kubernetes.io/projected/9a06cbe8-94ad-4a13-b29a-500768bb42fa-kube-api-access-pmp29\") pod \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.880408 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-catalog-content\") pod \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.880852 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-utilities\") pod \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\" (UID: \"9a06cbe8-94ad-4a13-b29a-500768bb42fa\") " Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.882503 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-utilities" (OuterVolumeSpecName: "utilities") pod "9a06cbe8-94ad-4a13-b29a-500768bb42fa" (UID: "9a06cbe8-94ad-4a13-b29a-500768bb42fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.882881 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.888072 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a06cbe8-94ad-4a13-b29a-500768bb42fa-kube-api-access-pmp29" (OuterVolumeSpecName: "kube-api-access-pmp29") pod "9a06cbe8-94ad-4a13-b29a-500768bb42fa" (UID: "9a06cbe8-94ad-4a13-b29a-500768bb42fa"). InnerVolumeSpecName "kube-api-access-pmp29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.946039 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a06cbe8-94ad-4a13-b29a-500768bb42fa" (UID: "9a06cbe8-94ad-4a13-b29a-500768bb42fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.985617 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmp29\" (UniqueName: \"kubernetes.io/projected/9a06cbe8-94ad-4a13-b29a-500768bb42fa-kube-api-access-pmp29\") on node \"crc\" DevicePath \"\"" Oct 10 16:37:09 crc kubenswrapper[4788]: I1010 16:37:09.985675 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a06cbe8-94ad-4a13-b29a-500768bb42fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.183419 4788 generic.go:334] "Generic (PLEG): container finished" podID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerID="23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e" exitCode=0 Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.183837 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4zllk" event={"ID":"9a06cbe8-94ad-4a13-b29a-500768bb42fa","Type":"ContainerDied","Data":"23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e"} Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.183879 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4zllk" event={"ID":"9a06cbe8-94ad-4a13-b29a-500768bb42fa","Type":"ContainerDied","Data":"df3b29d71298d4f75a7a57350d8e342976d65065e340f3681c01afcc3fee57f4"} Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.183902 4788 scope.go:117] "RemoveContainer" containerID="23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.184094 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4zllk" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.222852 4788 scope.go:117] "RemoveContainer" containerID="fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.254354 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4zllk"] Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.257281 4788 scope.go:117] "RemoveContainer" containerID="95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.265057 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4zllk"] Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.333793 4788 scope.go:117] "RemoveContainer" containerID="23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e" Oct 10 16:37:10 crc kubenswrapper[4788]: E1010 16:37:10.334505 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e\": container with ID starting with 23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e not found: ID does not exist" containerID="23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.334575 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e"} err="failed to get container status \"23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e\": rpc error: code = NotFound desc = could not find container \"23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e\": container with ID starting with 23745f26a71c173cc5ae32f0630b1d6bc8614a61a43179ec093e737b38b6223e not found: ID does not exist" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.334620 4788 scope.go:117] "RemoveContainer" containerID="fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b" Oct 10 16:37:10 crc kubenswrapper[4788]: E1010 16:37:10.335332 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b\": container with ID starting with fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b not found: ID does not exist" containerID="fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.335404 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b"} err="failed to get container status \"fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b\": rpc error: code = NotFound desc = could not find container \"fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b\": container with ID starting with fc2e9c0b367c720c7ac5434c9729a156f9adefd895898c88646bb000d20b920b not found: ID does not exist" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.335455 4788 scope.go:117] "RemoveContainer" containerID="95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f" Oct 10 16:37:10 crc kubenswrapper[4788]: E1010 16:37:10.336072 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f\": container with ID starting with 95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f not found: ID does not exist" containerID="95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f" Oct 10 16:37:10 crc kubenswrapper[4788]: I1010 16:37:10.336118 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f"} err="failed to get container status \"95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f\": rpc error: code = NotFound desc = could not find container \"95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f\": container with ID starting with 95b6709461585ff729a9f42fd8a9dcd82a533b7a08f4e127f770012252f83a2f not found: ID does not exist" Oct 10 16:37:12 crc kubenswrapper[4788]: I1010 16:37:12.250114 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" path="/var/lib/kubelet/pods/9a06cbe8-94ad-4a13-b29a-500768bb42fa/volumes" Oct 10 16:37:29 crc kubenswrapper[4788]: I1010 16:37:29.405748 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:37:29 crc kubenswrapper[4788]: I1010 16:37:29.406577 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:37:36 crc kubenswrapper[4788]: I1010 16:37:36.050772 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-wbld9"] Oct 10 16:37:36 crc kubenswrapper[4788]: I1010 16:37:36.058510 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-wbld9"] Oct 10 16:37:36 crc kubenswrapper[4788]: I1010 16:37:36.262400 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6066b1c7-e0b7-48a8-abd5-2d402169d6f7" path="/var/lib/kubelet/pods/6066b1c7-e0b7-48a8-abd5-2d402169d6f7/volumes" Oct 10 16:37:47 crc kubenswrapper[4788]: I1010 16:37:47.052441 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-a27e-account-create-6lzql"] Oct 10 16:37:47 crc kubenswrapper[4788]: I1010 16:37:47.065746 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-a27e-account-create-6lzql"] Oct 10 16:37:48 crc kubenswrapper[4788]: I1010 16:37:48.258094 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93f69a84-4f21-4331-afdc-657f573a2039" path="/var/lib/kubelet/pods/93f69a84-4f21-4331-afdc-657f573a2039/volumes" Oct 10 16:37:59 crc kubenswrapper[4788]: I1010 16:37:59.406256 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:37:59 crc kubenswrapper[4788]: I1010 16:37:59.406734 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:38:03 crc kubenswrapper[4788]: I1010 16:38:03.029358 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-mg4lr"] Oct 10 16:38:03 crc kubenswrapper[4788]: I1010 16:38:03.046736 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-mg4lr"] Oct 10 16:38:04 crc kubenswrapper[4788]: I1010 16:38:04.244953 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="842a80ee-3ba8-48ed-9513-dd197f1e2cd8" path="/var/lib/kubelet/pods/842a80ee-3ba8-48ed-9513-dd197f1e2cd8/volumes" Oct 10 16:38:18 crc kubenswrapper[4788]: I1010 16:38:18.864056 4788 scope.go:117] "RemoveContainer" containerID="71c05576423d8055d7b01451cc5e9c59286b40586cc0cdf9439a2466a5d48f67" Oct 10 16:38:18 crc kubenswrapper[4788]: I1010 16:38:18.933873 4788 scope.go:117] "RemoveContainer" containerID="13b813173dae0404d579256c1e5c19a9b9d8de0c3b91d7d04313fad2ee568443" Oct 10 16:38:18 crc kubenswrapper[4788]: I1010 16:38:18.982561 4788 scope.go:117] "RemoveContainer" containerID="17c9642df5822b0e0247784a9c5729443b55c55d19cacf2ae42cae42db60ec9b" Oct 10 16:38:29 crc kubenswrapper[4788]: I1010 16:38:29.406419 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:38:29 crc kubenswrapper[4788]: I1010 16:38:29.407000 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:38:29 crc kubenswrapper[4788]: I1010 16:38:29.407055 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:38:29 crc kubenswrapper[4788]: I1010 16:38:29.408013 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"55fb8770913a06b4815ab6ba7c27a4726c032dd815a83476336b3d0f44277d9b"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:38:29 crc kubenswrapper[4788]: I1010 16:38:29.408061 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://55fb8770913a06b4815ab6ba7c27a4726c032dd815a83476336b3d0f44277d9b" gracePeriod=600 Oct 10 16:38:30 crc kubenswrapper[4788]: I1010 16:38:30.149132 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="55fb8770913a06b4815ab6ba7c27a4726c032dd815a83476336b3d0f44277d9b" exitCode=0 Oct 10 16:38:30 crc kubenswrapper[4788]: I1010 16:38:30.149242 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"55fb8770913a06b4815ab6ba7c27a4726c032dd815a83476336b3d0f44277d9b"} Oct 10 16:38:30 crc kubenswrapper[4788]: I1010 16:38:30.149841 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe"} Oct 10 16:38:30 crc kubenswrapper[4788]: I1010 16:38:30.149874 4788 scope.go:117] "RemoveContainer" containerID="298a960328a3d56aec9aecec627fc67057fc33816ad446bad20fea20371082b3" Oct 10 16:39:19 crc kubenswrapper[4788]: I1010 16:39:19.116751 4788 scope.go:117] "RemoveContainer" containerID="dbca9f587ea5d14ae555bebda043de8949f825999129119d895fcb47c0f9308b" Oct 10 16:39:19 crc kubenswrapper[4788]: I1010 16:39:19.154631 4788 scope.go:117] "RemoveContainer" containerID="831743d71cc6a393fcb1d57c687e875a6a662d3ef468cdebaf209fa48e78486f" Oct 10 16:39:57 crc kubenswrapper[4788]: I1010 16:39:57.034566 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-sgtqs"] Oct 10 16:39:57 crc kubenswrapper[4788]: I1010 16:39:57.045977 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-sgtqs"] Oct 10 16:39:58 crc kubenswrapper[4788]: I1010 16:39:58.272987 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb4d0587-0164-4270-bb8d-93ec27bd3ab4" path="/var/lib/kubelet/pods/eb4d0587-0164-4270-bb8d-93ec27bd3ab4/volumes" Oct 10 16:40:08 crc kubenswrapper[4788]: I1010 16:40:08.030729 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-d613-account-create-99s9t"] Oct 10 16:40:08 crc kubenswrapper[4788]: I1010 16:40:08.040609 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-d613-account-create-99s9t"] Oct 10 16:40:08 crc kubenswrapper[4788]: I1010 16:40:08.245884 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c8a2701-d615-41c9-bf5a-ead64d030646" path="/var/lib/kubelet/pods/2c8a2701-d615-41c9-bf5a-ead64d030646/volumes" Oct 10 16:40:19 crc kubenswrapper[4788]: I1010 16:40:19.228951 4788 scope.go:117] "RemoveContainer" containerID="c5230e76d30353948a941e616ed001f193aa6353dfe70040b02810c412374d12" Oct 10 16:40:19 crc kubenswrapper[4788]: I1010 16:40:19.250779 4788 scope.go:117] "RemoveContainer" containerID="2cc1764388ca1ef85001d35b8a632684ea98a4e26a0a852d7a4042529fdd300c" Oct 10 16:40:22 crc kubenswrapper[4788]: I1010 16:40:22.050919 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-w4rcs"] Oct 10 16:40:22 crc kubenswrapper[4788]: I1010 16:40:22.066423 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-w4rcs"] Oct 10 16:40:22 crc kubenswrapper[4788]: I1010 16:40:22.250236 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d180f85a-fb1a-44ef-83f0-dae3057fb6e6" path="/var/lib/kubelet/pods/d180f85a-fb1a-44ef-83f0-dae3057fb6e6/volumes" Oct 10 16:40:29 crc kubenswrapper[4788]: I1010 16:40:29.406079 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:40:29 crc kubenswrapper[4788]: I1010 16:40:29.406695 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:40:45 crc kubenswrapper[4788]: I1010 16:40:45.051617 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-mk5cn"] Oct 10 16:40:45 crc kubenswrapper[4788]: I1010 16:40:45.060725 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-mk5cn"] Oct 10 16:40:46 crc kubenswrapper[4788]: I1010 16:40:46.250040 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a38bd59d-0d0f-4060-8a9f-09934dfbe079" path="/var/lib/kubelet/pods/a38bd59d-0d0f-4060-8a9f-09934dfbe079/volumes" Oct 10 16:40:54 crc kubenswrapper[4788]: I1010 16:40:54.038110 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-3e16-account-create-xzt2v"] Oct 10 16:40:54 crc kubenswrapper[4788]: I1010 16:40:54.050202 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-3e16-account-create-xzt2v"] Oct 10 16:40:54 crc kubenswrapper[4788]: I1010 16:40:54.250719 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="081f5e53-b6ea-41d5-ad7a-e1c11454af15" path="/var/lib/kubelet/pods/081f5e53-b6ea-41d5-ad7a-e1c11454af15/volumes" Oct 10 16:40:59 crc kubenswrapper[4788]: I1010 16:40:59.406893 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:40:59 crc kubenswrapper[4788]: I1010 16:40:59.407516 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:41:10 crc kubenswrapper[4788]: I1010 16:41:10.038681 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-pgqwd"] Oct 10 16:41:10 crc kubenswrapper[4788]: I1010 16:41:10.051436 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-pgqwd"] Oct 10 16:41:10 crc kubenswrapper[4788]: I1010 16:41:10.250553 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bc6dd5e-8a7b-4dd9-b766-c23d089eb081" path="/var/lib/kubelet/pods/3bc6dd5e-8a7b-4dd9-b766-c23d089eb081/volumes" Oct 10 16:41:19 crc kubenswrapper[4788]: I1010 16:41:19.386078 4788 scope.go:117] "RemoveContainer" containerID="6f698fb6c36499b5ad6e8f25c4ee094f2e76c24a0f36b7fdf9a1fc9869fc2816" Oct 10 16:41:19 crc kubenswrapper[4788]: I1010 16:41:19.428324 4788 scope.go:117] "RemoveContainer" containerID="935e4bc58e2a8cc3ce8c8917424ab739ab5a1e82e23f2013ef1253f5545fca4d" Oct 10 16:41:19 crc kubenswrapper[4788]: I1010 16:41:19.490775 4788 scope.go:117] "RemoveContainer" containerID="a1ca0c753209c8fa0d85457ef7e97700181df10af5766ce36f9810d1b8ba7b21" Oct 10 16:41:19 crc kubenswrapper[4788]: I1010 16:41:19.529408 4788 scope.go:117] "RemoveContainer" containerID="e8d98220b03bc653d3b269606c923c9f28f099265fd0051e87d118cbf6a3932c" Oct 10 16:41:29 crc kubenswrapper[4788]: I1010 16:41:29.406576 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:41:29 crc kubenswrapper[4788]: I1010 16:41:29.407293 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:41:29 crc kubenswrapper[4788]: I1010 16:41:29.407338 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:41:29 crc kubenswrapper[4788]: I1010 16:41:29.408181 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:41:29 crc kubenswrapper[4788]: I1010 16:41:29.408244 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" gracePeriod=600 Oct 10 16:41:29 crc kubenswrapper[4788]: E1010 16:41:29.537600 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:41:30 crc kubenswrapper[4788]: I1010 16:41:30.242485 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" exitCode=0 Oct 10 16:41:30 crc kubenswrapper[4788]: I1010 16:41:30.247156 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe"} Oct 10 16:41:30 crc kubenswrapper[4788]: I1010 16:41:30.247223 4788 scope.go:117] "RemoveContainer" containerID="55fb8770913a06b4815ab6ba7c27a4726c032dd815a83476336b3d0f44277d9b" Oct 10 16:41:30 crc kubenswrapper[4788]: I1010 16:41:30.247796 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:41:30 crc kubenswrapper[4788]: E1010 16:41:30.248070 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:41:44 crc kubenswrapper[4788]: I1010 16:41:44.234462 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:41:44 crc kubenswrapper[4788]: E1010 16:41:44.235538 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:41:55 crc kubenswrapper[4788]: I1010 16:41:55.235106 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:41:55 crc kubenswrapper[4788]: E1010 16:41:55.236886 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:42:09 crc kubenswrapper[4788]: I1010 16:42:09.234573 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:42:09 crc kubenswrapper[4788]: E1010 16:42:09.235641 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:42:22 crc kubenswrapper[4788]: I1010 16:42:22.235570 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:42:22 crc kubenswrapper[4788]: E1010 16:42:22.237565 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:42:34 crc kubenswrapper[4788]: I1010 16:42:34.235011 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:42:34 crc kubenswrapper[4788]: E1010 16:42:34.235894 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:42:47 crc kubenswrapper[4788]: I1010 16:42:47.234656 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:42:47 crc kubenswrapper[4788]: E1010 16:42:47.235489 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:42:59 crc kubenswrapper[4788]: I1010 16:42:59.234880 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:42:59 crc kubenswrapper[4788]: E1010 16:42:59.236082 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:43:13 crc kubenswrapper[4788]: I1010 16:43:13.233315 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:43:13 crc kubenswrapper[4788]: E1010 16:43:13.235034 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:43:26 crc kubenswrapper[4788]: I1010 16:43:26.235080 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:43:26 crc kubenswrapper[4788]: E1010 16:43:26.236270 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:43:39 crc kubenswrapper[4788]: I1010 16:43:39.236036 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:43:39 crc kubenswrapper[4788]: E1010 16:43:39.237492 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:43:50 crc kubenswrapper[4788]: I1010 16:43:50.237289 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:43:50 crc kubenswrapper[4788]: E1010 16:43:50.238518 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:44:04 crc kubenswrapper[4788]: I1010 16:44:04.234752 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:44:04 crc kubenswrapper[4788]: E1010 16:44:04.235849 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:44:18 crc kubenswrapper[4788]: I1010 16:44:18.239931 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:44:18 crc kubenswrapper[4788]: E1010 16:44:18.240746 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:44:24 crc kubenswrapper[4788]: I1010 16:44:24.428967 4788 generic.go:334] "Generic (PLEG): container finished" podID="6dbab995-d2d3-42cf-8f11-1484c7d00f4c" containerID="444d335fffb6499fdeefd39e412e1c7533c22be9568bc84f96fe066bba5923c5" exitCode=0 Oct 10 16:44:24 crc kubenswrapper[4788]: I1010 16:44:24.429200 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" event={"ID":"6dbab995-d2d3-42cf-8f11-1484c7d00f4c","Type":"ContainerDied","Data":"444d335fffb6499fdeefd39e412e1c7533c22be9568bc84f96fe066bba5923c5"} Oct 10 16:44:25 crc kubenswrapper[4788]: I1010 16:44:25.948696 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.022564 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-tripleo-cleanup-combined-ca-bundle\") pod \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.022608 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nfx7\" (UniqueName: \"kubernetes.io/projected/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-kube-api-access-8nfx7\") pod \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.022748 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ssh-key\") pod \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.022813 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ceph\") pod \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.023261 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-inventory\") pod \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\" (UID: \"6dbab995-d2d3-42cf-8f11-1484c7d00f4c\") " Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.029988 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "6dbab995-d2d3-42cf-8f11-1484c7d00f4c" (UID: "6dbab995-d2d3-42cf-8f11-1484c7d00f4c"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.030007 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ceph" (OuterVolumeSpecName: "ceph") pod "6dbab995-d2d3-42cf-8f11-1484c7d00f4c" (UID: "6dbab995-d2d3-42cf-8f11-1484c7d00f4c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.032963 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-kube-api-access-8nfx7" (OuterVolumeSpecName: "kube-api-access-8nfx7") pod "6dbab995-d2d3-42cf-8f11-1484c7d00f4c" (UID: "6dbab995-d2d3-42cf-8f11-1484c7d00f4c"). InnerVolumeSpecName "kube-api-access-8nfx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.065551 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-inventory" (OuterVolumeSpecName: "inventory") pod "6dbab995-d2d3-42cf-8f11-1484c7d00f4c" (UID: "6dbab995-d2d3-42cf-8f11-1484c7d00f4c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.077478 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6dbab995-d2d3-42cf-8f11-1484c7d00f4c" (UID: "6dbab995-d2d3-42cf-8f11-1484c7d00f4c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.127458 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.127529 4788 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.127544 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nfx7\" (UniqueName: \"kubernetes.io/projected/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-kube-api-access-8nfx7\") on node \"crc\" DevicePath \"\"" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.127556 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.127565 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6dbab995-d2d3-42cf-8f11-1484c7d00f4c-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.448429 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" event={"ID":"6dbab995-d2d3-42cf-8f11-1484c7d00f4c","Type":"ContainerDied","Data":"b85650119e85693ff750ba0b62e2dc04777d3b1d6d09981281f4a1ad367c5a35"} Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.448976 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b85650119e85693ff750ba0b62e2dc04777d3b1d6d09981281f4a1ad367c5a35" Oct 10 16:44:26 crc kubenswrapper[4788]: I1010 16:44:26.448484 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.051610 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-xsj2v"] Oct 10 16:44:30 crc kubenswrapper[4788]: E1010 16:44:30.053195 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerName="extract-utilities" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.053211 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerName="extract-utilities" Oct 10 16:44:30 crc kubenswrapper[4788]: E1010 16:44:30.053231 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dbab995-d2d3-42cf-8f11-1484c7d00f4c" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.053240 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dbab995-d2d3-42cf-8f11-1484c7d00f4c" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 10 16:44:30 crc kubenswrapper[4788]: E1010 16:44:30.053267 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerName="extract-content" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.053273 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerName="extract-content" Oct 10 16:44:30 crc kubenswrapper[4788]: E1010 16:44:30.053291 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerName="registry-server" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.053296 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerName="registry-server" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.053847 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a06cbe8-94ad-4a13-b29a-500768bb42fa" containerName="registry-server" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.053895 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dbab995-d2d3-42cf-8f11-1484c7d00f4c" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.055023 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.058956 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.061365 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-xsj2v"] Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.068065 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.068405 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.068567 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.142751 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.142855 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-inventory\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.142950 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.143036 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lftp\" (UniqueName: \"kubernetes.io/projected/171bf601-1ec1-4711-aacd-611751934a92-kube-api-access-9lftp\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.143061 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ceph\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.244622 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.244915 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lftp\" (UniqueName: \"kubernetes.io/projected/171bf601-1ec1-4711-aacd-611751934a92-kube-api-access-9lftp\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.244979 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ceph\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.245130 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.245281 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-inventory\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.249846 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-inventory\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.249946 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.259330 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ceph\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.262693 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.263539 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lftp\" (UniqueName: \"kubernetes.io/projected/171bf601-1ec1-4711-aacd-611751934a92-kube-api-access-9lftp\") pod \"bootstrap-openstack-openstack-cell1-xsj2v\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.385130 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.936918 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-xsj2v"] Oct 10 16:44:30 crc kubenswrapper[4788]: I1010 16:44:30.946339 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.348794 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dg7ml"] Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.351617 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.366352 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dg7ml"] Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.477613 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-utilities\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.477688 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-catalog-content\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.477717 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbhtr\" (UniqueName: \"kubernetes.io/projected/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-kube-api-access-nbhtr\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.518630 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" event={"ID":"171bf601-1ec1-4711-aacd-611751934a92","Type":"ContainerStarted","Data":"48bc9622a3452d819218ad9e5598c457b6ca5da1c58d8f34bcc837b844862405"} Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.518684 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" event={"ID":"171bf601-1ec1-4711-aacd-611751934a92","Type":"ContainerStarted","Data":"a95b17fe1e7319a80e3510506126633ec5fda7e4713c1ddeeca6488c4c836110"} Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.542021 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" podStartSLOduration=1.356510879 podStartE2EDuration="1.54199775s" podCreationTimestamp="2025-10-10 16:44:30 +0000 UTC" firstStartedPulling="2025-10-10 16:44:30.946117175 +0000 UTC m=+7173.395832723" lastFinishedPulling="2025-10-10 16:44:31.131604046 +0000 UTC m=+7173.581319594" observedRunningTime="2025-10-10 16:44:31.536831508 +0000 UTC m=+7173.986547056" watchObservedRunningTime="2025-10-10 16:44:31.54199775 +0000 UTC m=+7173.991713298" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.582606 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-catalog-content\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.582667 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbhtr\" (UniqueName: \"kubernetes.io/projected/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-kube-api-access-nbhtr\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.583061 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-utilities\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.583222 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-catalog-content\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.583552 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-utilities\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.600807 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbhtr\" (UniqueName: \"kubernetes.io/projected/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-kube-api-access-nbhtr\") pod \"certified-operators-dg7ml\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:31 crc kubenswrapper[4788]: I1010 16:44:31.686106 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:32 crc kubenswrapper[4788]: I1010 16:44:32.287253 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dg7ml"] Oct 10 16:44:32 crc kubenswrapper[4788]: I1010 16:44:32.540409 4788 generic.go:334] "Generic (PLEG): container finished" podID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerID="0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf" exitCode=0 Oct 10 16:44:32 crc kubenswrapper[4788]: I1010 16:44:32.541053 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dg7ml" event={"ID":"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f","Type":"ContainerDied","Data":"0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf"} Oct 10 16:44:32 crc kubenswrapper[4788]: I1010 16:44:32.541099 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dg7ml" event={"ID":"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f","Type":"ContainerStarted","Data":"bbcfe67aac85a5179038606f442a43e98681c57b3efa255a4650ebd39c618e48"} Oct 10 16:44:33 crc kubenswrapper[4788]: I1010 16:44:33.234578 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:44:33 crc kubenswrapper[4788]: E1010 16:44:33.235308 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:44:33 crc kubenswrapper[4788]: I1010 16:44:33.553320 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dg7ml" event={"ID":"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f","Type":"ContainerStarted","Data":"684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776"} Oct 10 16:44:34 crc kubenswrapper[4788]: I1010 16:44:34.565113 4788 generic.go:334] "Generic (PLEG): container finished" podID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerID="684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776" exitCode=0 Oct 10 16:44:34 crc kubenswrapper[4788]: I1010 16:44:34.565194 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dg7ml" event={"ID":"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f","Type":"ContainerDied","Data":"684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776"} Oct 10 16:44:35 crc kubenswrapper[4788]: I1010 16:44:35.619191 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dg7ml" event={"ID":"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f","Type":"ContainerStarted","Data":"58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c"} Oct 10 16:44:35 crc kubenswrapper[4788]: I1010 16:44:35.640470 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dg7ml" podStartSLOduration=2.203714747 podStartE2EDuration="4.640449187s" podCreationTimestamp="2025-10-10 16:44:31 +0000 UTC" firstStartedPulling="2025-10-10 16:44:32.541954755 +0000 UTC m=+7174.991670303" lastFinishedPulling="2025-10-10 16:44:34.978689195 +0000 UTC m=+7177.428404743" observedRunningTime="2025-10-10 16:44:35.640412096 +0000 UTC m=+7178.090127644" watchObservedRunningTime="2025-10-10 16:44:35.640449187 +0000 UTC m=+7178.090164735" Oct 10 16:44:41 crc kubenswrapper[4788]: I1010 16:44:41.686818 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:41 crc kubenswrapper[4788]: I1010 16:44:41.687797 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:41 crc kubenswrapper[4788]: I1010 16:44:41.752729 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:42 crc kubenswrapper[4788]: I1010 16:44:42.751263 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:42 crc kubenswrapper[4788]: I1010 16:44:42.834379 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dg7ml"] Oct 10 16:44:44 crc kubenswrapper[4788]: I1010 16:44:44.718724 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dg7ml" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerName="registry-server" containerID="cri-o://58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c" gracePeriod=2 Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.269642 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.374390 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-utilities\") pod \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.374964 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-catalog-content\") pod \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.375188 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbhtr\" (UniqueName: \"kubernetes.io/projected/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-kube-api-access-nbhtr\") pod \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\" (UID: \"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f\") " Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.376264 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-utilities" (OuterVolumeSpecName: "utilities") pod "f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" (UID: "f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.379125 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.388970 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-kube-api-access-nbhtr" (OuterVolumeSpecName: "kube-api-access-nbhtr") pod "f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" (UID: "f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f"). InnerVolumeSpecName "kube-api-access-nbhtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.434417 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" (UID: "f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.481079 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.481119 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbhtr\" (UniqueName: \"kubernetes.io/projected/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f-kube-api-access-nbhtr\") on node \"crc\" DevicePath \"\"" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.731454 4788 generic.go:334] "Generic (PLEG): container finished" podID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerID="58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c" exitCode=0 Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.731501 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dg7ml" event={"ID":"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f","Type":"ContainerDied","Data":"58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c"} Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.731539 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dg7ml" event={"ID":"f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f","Type":"ContainerDied","Data":"bbcfe67aac85a5179038606f442a43e98681c57b3efa255a4650ebd39c618e48"} Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.731572 4788 scope.go:117] "RemoveContainer" containerID="58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.731622 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dg7ml" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.774656 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dg7ml"] Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.784539 4788 scope.go:117] "RemoveContainer" containerID="684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.786264 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dg7ml"] Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.810859 4788 scope.go:117] "RemoveContainer" containerID="0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.878698 4788 scope.go:117] "RemoveContainer" containerID="58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c" Oct 10 16:44:45 crc kubenswrapper[4788]: E1010 16:44:45.879237 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c\": container with ID starting with 58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c not found: ID does not exist" containerID="58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.879293 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c"} err="failed to get container status \"58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c\": rpc error: code = NotFound desc = could not find container \"58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c\": container with ID starting with 58cf3c5e219079ca312496a368de57fab684d868053fa66414059f055e97583c not found: ID does not exist" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.879318 4788 scope.go:117] "RemoveContainer" containerID="684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776" Oct 10 16:44:45 crc kubenswrapper[4788]: E1010 16:44:45.879739 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776\": container with ID starting with 684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776 not found: ID does not exist" containerID="684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.879800 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776"} err="failed to get container status \"684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776\": rpc error: code = NotFound desc = could not find container \"684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776\": container with ID starting with 684956ed7392159d89ee0662d2c007fb86f64b739674815d8cf6332d7ad9a776 not found: ID does not exist" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.879854 4788 scope.go:117] "RemoveContainer" containerID="0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf" Oct 10 16:44:45 crc kubenswrapper[4788]: E1010 16:44:45.880250 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf\": container with ID starting with 0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf not found: ID does not exist" containerID="0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf" Oct 10 16:44:45 crc kubenswrapper[4788]: I1010 16:44:45.880277 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf"} err="failed to get container status \"0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf\": rpc error: code = NotFound desc = could not find container \"0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf\": container with ID starting with 0e6bcdfc8d8a6cda0f7feb27f98bf074d08b5436a2eb24664fea46b661298eaf not found: ID does not exist" Oct 10 16:44:46 crc kubenswrapper[4788]: I1010 16:44:46.252308 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" path="/var/lib/kubelet/pods/f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f/volumes" Oct 10 16:44:47 crc kubenswrapper[4788]: I1010 16:44:47.234533 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:44:47 crc kubenswrapper[4788]: E1010 16:44:47.235130 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.179197 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4"] Oct 10 16:45:00 crc kubenswrapper[4788]: E1010 16:45:00.180449 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerName="registry-server" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.180468 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerName="registry-server" Oct 10 16:45:00 crc kubenswrapper[4788]: E1010 16:45:00.180492 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerName="extract-content" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.180499 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerName="extract-content" Oct 10 16:45:00 crc kubenswrapper[4788]: E1010 16:45:00.180529 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerName="extract-utilities" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.180538 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerName="extract-utilities" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.180855 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8edf5f1-8bb7-44e4-a2ba-c4576cbff47f" containerName="registry-server" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.187153 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.189597 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.193472 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.203079 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4"] Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.243588 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:45:00 crc kubenswrapper[4788]: E1010 16:45:00.244260 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.365722 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-secret-volume\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.365801 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85cs5\" (UniqueName: \"kubernetes.io/projected/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-kube-api-access-85cs5\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.365831 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-config-volume\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.468647 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-secret-volume\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.469106 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85cs5\" (UniqueName: \"kubernetes.io/projected/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-kube-api-access-85cs5\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.469130 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-config-volume\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.470042 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-config-volume\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.476279 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-secret-volume\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.491741 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85cs5\" (UniqueName: \"kubernetes.io/projected/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-kube-api-access-85cs5\") pod \"collect-profiles-29335245-p25n4\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:00 crc kubenswrapper[4788]: I1010 16:45:00.554865 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:01 crc kubenswrapper[4788]: I1010 16:45:01.074840 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4"] Oct 10 16:45:01 crc kubenswrapper[4788]: I1010 16:45:01.910412 4788 generic.go:334] "Generic (PLEG): container finished" podID="523d2a02-011c-4feb-a2e5-8468d7f1e3eb" containerID="92a5753885667dbaaa8ca6ebaf4aada52c46271994bfe911f64568ce6a6c08f3" exitCode=0 Oct 10 16:45:01 crc kubenswrapper[4788]: I1010 16:45:01.910469 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" event={"ID":"523d2a02-011c-4feb-a2e5-8468d7f1e3eb","Type":"ContainerDied","Data":"92a5753885667dbaaa8ca6ebaf4aada52c46271994bfe911f64568ce6a6c08f3"} Oct 10 16:45:01 crc kubenswrapper[4788]: I1010 16:45:01.910805 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" event={"ID":"523d2a02-011c-4feb-a2e5-8468d7f1e3eb","Type":"ContainerStarted","Data":"2b32b7eca2ff72ae0bc15456fef8e4450a5443ffd92d3ff756b31d3999224cae"} Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.341684 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.541850 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-secret-volume\") pod \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.541939 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-config-volume\") pod \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.542029 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85cs5\" (UniqueName: \"kubernetes.io/projected/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-kube-api-access-85cs5\") pod \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\" (UID: \"523d2a02-011c-4feb-a2e5-8468d7f1e3eb\") " Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.542773 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "523d2a02-011c-4feb-a2e5-8468d7f1e3eb" (UID: "523d2a02-011c-4feb-a2e5-8468d7f1e3eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.558377 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "523d2a02-011c-4feb-a2e5-8468d7f1e3eb" (UID: "523d2a02-011c-4feb-a2e5-8468d7f1e3eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.558440 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-kube-api-access-85cs5" (OuterVolumeSpecName: "kube-api-access-85cs5") pod "523d2a02-011c-4feb-a2e5-8468d7f1e3eb" (UID: "523d2a02-011c-4feb-a2e5-8468d7f1e3eb"). InnerVolumeSpecName "kube-api-access-85cs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.645460 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.645517 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.645538 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85cs5\" (UniqueName: \"kubernetes.io/projected/523d2a02-011c-4feb-a2e5-8468d7f1e3eb-kube-api-access-85cs5\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.931426 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" event={"ID":"523d2a02-011c-4feb-a2e5-8468d7f1e3eb","Type":"ContainerDied","Data":"2b32b7eca2ff72ae0bc15456fef8e4450a5443ffd92d3ff756b31d3999224cae"} Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.931468 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b32b7eca2ff72ae0bc15456fef8e4450a5443ffd92d3ff756b31d3999224cae" Oct 10 16:45:03 crc kubenswrapper[4788]: I1010 16:45:03.931536 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4" Oct 10 16:45:04 crc kubenswrapper[4788]: I1010 16:45:04.421337 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw"] Oct 10 16:45:04 crc kubenswrapper[4788]: I1010 16:45:04.429648 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335200-pbnlw"] Oct 10 16:45:06 crc kubenswrapper[4788]: I1010 16:45:06.254411 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="035cfcc7-90ca-4547-b3b7-2094e8ff225e" path="/var/lib/kubelet/pods/035cfcc7-90ca-4547-b3b7-2094e8ff225e/volumes" Oct 10 16:45:12 crc kubenswrapper[4788]: I1010 16:45:12.234811 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:45:12 crc kubenswrapper[4788]: E1010 16:45:12.235563 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.214077 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2whlk"] Oct 10 16:45:16 crc kubenswrapper[4788]: E1010 16:45:16.215495 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="523d2a02-011c-4feb-a2e5-8468d7f1e3eb" containerName="collect-profiles" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.215514 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="523d2a02-011c-4feb-a2e5-8468d7f1e3eb" containerName="collect-profiles" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.215846 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="523d2a02-011c-4feb-a2e5-8468d7f1e3eb" containerName="collect-profiles" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.217880 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.228377 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2whlk"] Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.379244 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-catalog-content\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.379602 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gpmf\" (UniqueName: \"kubernetes.io/projected/6b5b3d53-47a9-468c-8b29-905ca596859b-kube-api-access-9gpmf\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.380038 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-utilities\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.483922 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-catalog-content\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.484024 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gpmf\" (UniqueName: \"kubernetes.io/projected/6b5b3d53-47a9-468c-8b29-905ca596859b-kube-api-access-9gpmf\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.484172 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-utilities\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.484658 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-utilities\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.484852 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-catalog-content\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.530210 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gpmf\" (UniqueName: \"kubernetes.io/projected/6b5b3d53-47a9-468c-8b29-905ca596859b-kube-api-access-9gpmf\") pod \"redhat-marketplace-2whlk\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:16 crc kubenswrapper[4788]: I1010 16:45:16.556281 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:17 crc kubenswrapper[4788]: I1010 16:45:17.089495 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2whlk"] Oct 10 16:45:18 crc kubenswrapper[4788]: I1010 16:45:18.105316 4788 generic.go:334] "Generic (PLEG): container finished" podID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerID="df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a" exitCode=0 Oct 10 16:45:18 crc kubenswrapper[4788]: I1010 16:45:18.105430 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2whlk" event={"ID":"6b5b3d53-47a9-468c-8b29-905ca596859b","Type":"ContainerDied","Data":"df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a"} Oct 10 16:45:18 crc kubenswrapper[4788]: I1010 16:45:18.105649 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2whlk" event={"ID":"6b5b3d53-47a9-468c-8b29-905ca596859b","Type":"ContainerStarted","Data":"83ebffb186eeb2c0d6005b9427982f38decbfc4c4757003816eaeacfe54d5791"} Oct 10 16:45:19 crc kubenswrapper[4788]: I1010 16:45:19.115715 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2whlk" event={"ID":"6b5b3d53-47a9-468c-8b29-905ca596859b","Type":"ContainerStarted","Data":"f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1"} Oct 10 16:45:19 crc kubenswrapper[4788]: I1010 16:45:19.787004 4788 scope.go:117] "RemoveContainer" containerID="9108b8823a0aeda1eeb999175ae7707a71c827e6afc666b06281dd814e69d626" Oct 10 16:45:20 crc kubenswrapper[4788]: I1010 16:45:20.128587 4788 generic.go:334] "Generic (PLEG): container finished" podID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerID="f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1" exitCode=0 Oct 10 16:45:20 crc kubenswrapper[4788]: I1010 16:45:20.128718 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2whlk" event={"ID":"6b5b3d53-47a9-468c-8b29-905ca596859b","Type":"ContainerDied","Data":"f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1"} Oct 10 16:45:21 crc kubenswrapper[4788]: I1010 16:45:21.143425 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2whlk" event={"ID":"6b5b3d53-47a9-468c-8b29-905ca596859b","Type":"ContainerStarted","Data":"91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e"} Oct 10 16:45:21 crc kubenswrapper[4788]: I1010 16:45:21.172244 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2whlk" podStartSLOduration=2.724855902 podStartE2EDuration="5.172219893s" podCreationTimestamp="2025-10-10 16:45:16 +0000 UTC" firstStartedPulling="2025-10-10 16:45:18.108579908 +0000 UTC m=+7220.558295466" lastFinishedPulling="2025-10-10 16:45:20.555943909 +0000 UTC m=+7223.005659457" observedRunningTime="2025-10-10 16:45:21.161128979 +0000 UTC m=+7223.610844527" watchObservedRunningTime="2025-10-10 16:45:21.172219893 +0000 UTC m=+7223.621935431" Oct 10 16:45:26 crc kubenswrapper[4788]: I1010 16:45:26.556971 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:26 crc kubenswrapper[4788]: I1010 16:45:26.557611 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:26 crc kubenswrapper[4788]: I1010 16:45:26.636255 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:27 crc kubenswrapper[4788]: I1010 16:45:27.234991 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:45:27 crc kubenswrapper[4788]: E1010 16:45:27.235565 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:45:27 crc kubenswrapper[4788]: I1010 16:45:27.242754 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:27 crc kubenswrapper[4788]: I1010 16:45:27.283432 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2whlk"] Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.213480 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2whlk" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerName="registry-server" containerID="cri-o://91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e" gracePeriod=2 Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.684695 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.779950 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gpmf\" (UniqueName: \"kubernetes.io/projected/6b5b3d53-47a9-468c-8b29-905ca596859b-kube-api-access-9gpmf\") pod \"6b5b3d53-47a9-468c-8b29-905ca596859b\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.780033 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-utilities\") pod \"6b5b3d53-47a9-468c-8b29-905ca596859b\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.780267 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-catalog-content\") pod \"6b5b3d53-47a9-468c-8b29-905ca596859b\" (UID: \"6b5b3d53-47a9-468c-8b29-905ca596859b\") " Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.780939 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-utilities" (OuterVolumeSpecName: "utilities") pod "6b5b3d53-47a9-468c-8b29-905ca596859b" (UID: "6b5b3d53-47a9-468c-8b29-905ca596859b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.785480 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b5b3d53-47a9-468c-8b29-905ca596859b-kube-api-access-9gpmf" (OuterVolumeSpecName: "kube-api-access-9gpmf") pod "6b5b3d53-47a9-468c-8b29-905ca596859b" (UID: "6b5b3d53-47a9-468c-8b29-905ca596859b"). InnerVolumeSpecName "kube-api-access-9gpmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.793031 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b5b3d53-47a9-468c-8b29-905ca596859b" (UID: "6b5b3d53-47a9-468c-8b29-905ca596859b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.882843 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.882889 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gpmf\" (UniqueName: \"kubernetes.io/projected/6b5b3d53-47a9-468c-8b29-905ca596859b-kube-api-access-9gpmf\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:29 crc kubenswrapper[4788]: I1010 16:45:29.882907 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b5b3d53-47a9-468c-8b29-905ca596859b-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.224099 4788 generic.go:334] "Generic (PLEG): container finished" podID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerID="91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e" exitCode=0 Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.224166 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2whlk" event={"ID":"6b5b3d53-47a9-468c-8b29-905ca596859b","Type":"ContainerDied","Data":"91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e"} Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.224472 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2whlk" event={"ID":"6b5b3d53-47a9-468c-8b29-905ca596859b","Type":"ContainerDied","Data":"83ebffb186eeb2c0d6005b9427982f38decbfc4c4757003816eaeacfe54d5791"} Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.224497 4788 scope.go:117] "RemoveContainer" containerID="91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.224191 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2whlk" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.264616 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2whlk"] Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.265157 4788 scope.go:117] "RemoveContainer" containerID="f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.272232 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2whlk"] Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.298248 4788 scope.go:117] "RemoveContainer" containerID="df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.341017 4788 scope.go:117] "RemoveContainer" containerID="91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e" Oct 10 16:45:30 crc kubenswrapper[4788]: E1010 16:45:30.341938 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e\": container with ID starting with 91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e not found: ID does not exist" containerID="91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.341998 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e"} err="failed to get container status \"91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e\": rpc error: code = NotFound desc = could not find container \"91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e\": container with ID starting with 91584d67e6b47b6290d96c171a91caaa4216e253d747f093cfc674904d28ce5e not found: ID does not exist" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.342032 4788 scope.go:117] "RemoveContainer" containerID="f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1" Oct 10 16:45:30 crc kubenswrapper[4788]: E1010 16:45:30.342749 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1\": container with ID starting with f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1 not found: ID does not exist" containerID="f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.342860 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1"} err="failed to get container status \"f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1\": rpc error: code = NotFound desc = could not find container \"f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1\": container with ID starting with f4740dab5fafb08f5f963cad5c020bd3081e8e357f1afcd30aca98481b6a55b1 not found: ID does not exist" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.342933 4788 scope.go:117] "RemoveContainer" containerID="df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a" Oct 10 16:45:30 crc kubenswrapper[4788]: E1010 16:45:30.343544 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a\": container with ID starting with df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a not found: ID does not exist" containerID="df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a" Oct 10 16:45:30 crc kubenswrapper[4788]: I1010 16:45:30.343583 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a"} err="failed to get container status \"df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a\": rpc error: code = NotFound desc = could not find container \"df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a\": container with ID starting with df16f3815e747231a2d7e580438d728a6208e4bc1ab070f5841e797eebf4722a not found: ID does not exist" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.251632 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" path="/var/lib/kubelet/pods/6b5b3d53-47a9-468c-8b29-905ca596859b/volumes" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.718790 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ntr52"] Oct 10 16:45:32 crc kubenswrapper[4788]: E1010 16:45:32.719334 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerName="registry-server" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.719355 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerName="registry-server" Oct 10 16:45:32 crc kubenswrapper[4788]: E1010 16:45:32.719384 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerName="extract-content" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.719390 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerName="extract-content" Oct 10 16:45:32 crc kubenswrapper[4788]: E1010 16:45:32.719443 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerName="extract-utilities" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.719451 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerName="extract-utilities" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.719711 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5b3d53-47a9-468c-8b29-905ca596859b" containerName="registry-server" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.721381 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.732501 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ntr52"] Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.761366 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-utilities\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.761515 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtxsp\" (UniqueName: \"kubernetes.io/projected/921b31bc-b66c-4727-8276-cafb89e8868a-kube-api-access-wtxsp\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.761644 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-catalog-content\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.864380 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-utilities\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.864563 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtxsp\" (UniqueName: \"kubernetes.io/projected/921b31bc-b66c-4727-8276-cafb89e8868a-kube-api-access-wtxsp\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.864719 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-catalog-content\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.865545 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-catalog-content\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.865689 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-utilities\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:32 crc kubenswrapper[4788]: I1010 16:45:32.893026 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtxsp\" (UniqueName: \"kubernetes.io/projected/921b31bc-b66c-4727-8276-cafb89e8868a-kube-api-access-wtxsp\") pod \"redhat-operators-ntr52\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:33 crc kubenswrapper[4788]: I1010 16:45:33.053678 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:33 crc kubenswrapper[4788]: I1010 16:45:33.554314 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ntr52"] Oct 10 16:45:34 crc kubenswrapper[4788]: I1010 16:45:34.284093 4788 generic.go:334] "Generic (PLEG): container finished" podID="921b31bc-b66c-4727-8276-cafb89e8868a" containerID="eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705" exitCode=0 Oct 10 16:45:34 crc kubenswrapper[4788]: I1010 16:45:34.284164 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntr52" event={"ID":"921b31bc-b66c-4727-8276-cafb89e8868a","Type":"ContainerDied","Data":"eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705"} Oct 10 16:45:34 crc kubenswrapper[4788]: I1010 16:45:34.284436 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntr52" event={"ID":"921b31bc-b66c-4727-8276-cafb89e8868a","Type":"ContainerStarted","Data":"9158e6c2140b9b15050439ee028a3c4195288cc82fd26fe7fde509438e8d7696"} Oct 10 16:45:35 crc kubenswrapper[4788]: I1010 16:45:35.298759 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntr52" event={"ID":"921b31bc-b66c-4727-8276-cafb89e8868a","Type":"ContainerStarted","Data":"aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0"} Oct 10 16:45:38 crc kubenswrapper[4788]: I1010 16:45:38.244560 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:45:38 crc kubenswrapper[4788]: E1010 16:45:38.245216 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:45:39 crc kubenswrapper[4788]: I1010 16:45:39.342909 4788 generic.go:334] "Generic (PLEG): container finished" podID="921b31bc-b66c-4727-8276-cafb89e8868a" containerID="aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0" exitCode=0 Oct 10 16:45:39 crc kubenswrapper[4788]: I1010 16:45:39.343082 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntr52" event={"ID":"921b31bc-b66c-4727-8276-cafb89e8868a","Type":"ContainerDied","Data":"aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0"} Oct 10 16:45:40 crc kubenswrapper[4788]: I1010 16:45:40.355502 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntr52" event={"ID":"921b31bc-b66c-4727-8276-cafb89e8868a","Type":"ContainerStarted","Data":"755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466"} Oct 10 16:45:40 crc kubenswrapper[4788]: I1010 16:45:40.387776 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ntr52" podStartSLOduration=2.879635233 podStartE2EDuration="8.387752402s" podCreationTimestamp="2025-10-10 16:45:32 +0000 UTC" firstStartedPulling="2025-10-10 16:45:34.285858686 +0000 UTC m=+7236.735574234" lastFinishedPulling="2025-10-10 16:45:39.793975865 +0000 UTC m=+7242.243691403" observedRunningTime="2025-10-10 16:45:40.375519566 +0000 UTC m=+7242.825235124" watchObservedRunningTime="2025-10-10 16:45:40.387752402 +0000 UTC m=+7242.837467960" Oct 10 16:45:43 crc kubenswrapper[4788]: I1010 16:45:43.054194 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:43 crc kubenswrapper[4788]: I1010 16:45:43.054488 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:44 crc kubenswrapper[4788]: I1010 16:45:44.105350 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ntr52" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="registry-server" probeResult="failure" output=< Oct 10 16:45:44 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 16:45:44 crc kubenswrapper[4788]: > Oct 10 16:45:51 crc kubenswrapper[4788]: I1010 16:45:51.234657 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:45:51 crc kubenswrapper[4788]: E1010 16:45:51.235540 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:45:53 crc kubenswrapper[4788]: I1010 16:45:53.128690 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:53 crc kubenswrapper[4788]: I1010 16:45:53.192726 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:54 crc kubenswrapper[4788]: I1010 16:45:54.367579 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ntr52"] Oct 10 16:45:54 crc kubenswrapper[4788]: I1010 16:45:54.499416 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ntr52" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="registry-server" containerID="cri-o://755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466" gracePeriod=2 Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.207844 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.292714 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-utilities\") pod \"921b31bc-b66c-4727-8276-cafb89e8868a\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.292978 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtxsp\" (UniqueName: \"kubernetes.io/projected/921b31bc-b66c-4727-8276-cafb89e8868a-kube-api-access-wtxsp\") pod \"921b31bc-b66c-4727-8276-cafb89e8868a\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.293033 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-catalog-content\") pod \"921b31bc-b66c-4727-8276-cafb89e8868a\" (UID: \"921b31bc-b66c-4727-8276-cafb89e8868a\") " Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.294575 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-utilities" (OuterVolumeSpecName: "utilities") pod "921b31bc-b66c-4727-8276-cafb89e8868a" (UID: "921b31bc-b66c-4727-8276-cafb89e8868a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.299935 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/921b31bc-b66c-4727-8276-cafb89e8868a-kube-api-access-wtxsp" (OuterVolumeSpecName: "kube-api-access-wtxsp") pod "921b31bc-b66c-4727-8276-cafb89e8868a" (UID: "921b31bc-b66c-4727-8276-cafb89e8868a"). InnerVolumeSpecName "kube-api-access-wtxsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.381959 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "921b31bc-b66c-4727-8276-cafb89e8868a" (UID: "921b31bc-b66c-4727-8276-cafb89e8868a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.394690 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtxsp\" (UniqueName: \"kubernetes.io/projected/921b31bc-b66c-4727-8276-cafb89e8868a-kube-api-access-wtxsp\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.394719 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.394730 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/921b31bc-b66c-4727-8276-cafb89e8868a-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.512453 4788 generic.go:334] "Generic (PLEG): container finished" podID="921b31bc-b66c-4727-8276-cafb89e8868a" containerID="755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466" exitCode=0 Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.512507 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntr52" event={"ID":"921b31bc-b66c-4727-8276-cafb89e8868a","Type":"ContainerDied","Data":"755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466"} Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.512573 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ntr52" event={"ID":"921b31bc-b66c-4727-8276-cafb89e8868a","Type":"ContainerDied","Data":"9158e6c2140b9b15050439ee028a3c4195288cc82fd26fe7fde509438e8d7696"} Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.512596 4788 scope.go:117] "RemoveContainer" containerID="755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.513087 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ntr52" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.566393 4788 scope.go:117] "RemoveContainer" containerID="aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.570469 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ntr52"] Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.586194 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ntr52"] Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.596205 4788 scope.go:117] "RemoveContainer" containerID="eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.651969 4788 scope.go:117] "RemoveContainer" containerID="755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466" Oct 10 16:45:55 crc kubenswrapper[4788]: E1010 16:45:55.652553 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466\": container with ID starting with 755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466 not found: ID does not exist" containerID="755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.652641 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466"} err="failed to get container status \"755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466\": rpc error: code = NotFound desc = could not find container \"755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466\": container with ID starting with 755a2c318f455a7dd08a5b85fa07bac3edf507da8616ce06c980ac3f73984466 not found: ID does not exist" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.652684 4788 scope.go:117] "RemoveContainer" containerID="aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0" Oct 10 16:45:55 crc kubenswrapper[4788]: E1010 16:45:55.653122 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0\": container with ID starting with aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0 not found: ID does not exist" containerID="aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.653248 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0"} err="failed to get container status \"aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0\": rpc error: code = NotFound desc = could not find container \"aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0\": container with ID starting with aa87923cb63bc826d7648ae3ce424c0a4d600c3b8287344f843417064ead85b0 not found: ID does not exist" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.653330 4788 scope.go:117] "RemoveContainer" containerID="eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705" Oct 10 16:45:55 crc kubenswrapper[4788]: E1010 16:45:55.653641 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705\": container with ID starting with eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705 not found: ID does not exist" containerID="eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705" Oct 10 16:45:55 crc kubenswrapper[4788]: I1010 16:45:55.653693 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705"} err="failed to get container status \"eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705\": rpc error: code = NotFound desc = could not find container \"eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705\": container with ID starting with eadf8168c46e50ab670a40dac10232307c7fba9b7880823b9a7463fd88370705 not found: ID does not exist" Oct 10 16:45:56 crc kubenswrapper[4788]: I1010 16:45:56.257402 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" path="/var/lib/kubelet/pods/921b31bc-b66c-4727-8276-cafb89e8868a/volumes" Oct 10 16:46:06 crc kubenswrapper[4788]: I1010 16:46:06.234214 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:46:06 crc kubenswrapper[4788]: E1010 16:46:06.235478 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:46:17 crc kubenswrapper[4788]: I1010 16:46:17.234100 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:46:17 crc kubenswrapper[4788]: E1010 16:46:17.235548 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:46:32 crc kubenswrapper[4788]: I1010 16:46:32.234687 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:46:32 crc kubenswrapper[4788]: I1010 16:46:32.936839 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"e843e596e0077c58474a49e15e5d3b91c305413c7ff87925b3f77a716e6e9ec3"} Oct 10 16:47:36 crc kubenswrapper[4788]: I1010 16:47:36.664769 4788 generic.go:334] "Generic (PLEG): container finished" podID="171bf601-1ec1-4711-aacd-611751934a92" containerID="48bc9622a3452d819218ad9e5598c457b6ca5da1c58d8f34bcc837b844862405" exitCode=0 Oct 10 16:47:36 crc kubenswrapper[4788]: I1010 16:47:36.664910 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" event={"ID":"171bf601-1ec1-4711-aacd-611751934a92","Type":"ContainerDied","Data":"48bc9622a3452d819218ad9e5598c457b6ca5da1c58d8f34bcc837b844862405"} Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.278898 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.431206 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-bootstrap-combined-ca-bundle\") pod \"171bf601-1ec1-4711-aacd-611751934a92\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.431415 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ceph\") pod \"171bf601-1ec1-4711-aacd-611751934a92\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.431548 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ssh-key\") pod \"171bf601-1ec1-4711-aacd-611751934a92\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.431594 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-inventory\") pod \"171bf601-1ec1-4711-aacd-611751934a92\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.431632 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lftp\" (UniqueName: \"kubernetes.io/projected/171bf601-1ec1-4711-aacd-611751934a92-kube-api-access-9lftp\") pod \"171bf601-1ec1-4711-aacd-611751934a92\" (UID: \"171bf601-1ec1-4711-aacd-611751934a92\") " Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.436610 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "171bf601-1ec1-4711-aacd-611751934a92" (UID: "171bf601-1ec1-4711-aacd-611751934a92"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.444055 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/171bf601-1ec1-4711-aacd-611751934a92-kube-api-access-9lftp" (OuterVolumeSpecName: "kube-api-access-9lftp") pod "171bf601-1ec1-4711-aacd-611751934a92" (UID: "171bf601-1ec1-4711-aacd-611751934a92"). InnerVolumeSpecName "kube-api-access-9lftp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.444279 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ceph" (OuterVolumeSpecName: "ceph") pod "171bf601-1ec1-4711-aacd-611751934a92" (UID: "171bf601-1ec1-4711-aacd-611751934a92"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.469130 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-inventory" (OuterVolumeSpecName: "inventory") pod "171bf601-1ec1-4711-aacd-611751934a92" (UID: "171bf601-1ec1-4711-aacd-611751934a92"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.474966 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "171bf601-1ec1-4711-aacd-611751934a92" (UID: "171bf601-1ec1-4711-aacd-611751934a92"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.534047 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.534084 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.534093 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.534104 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lftp\" (UniqueName: \"kubernetes.io/projected/171bf601-1ec1-4711-aacd-611751934a92-kube-api-access-9lftp\") on node \"crc\" DevicePath \"\"" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.534114 4788 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/171bf601-1ec1-4711-aacd-611751934a92-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.697617 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" event={"ID":"171bf601-1ec1-4711-aacd-611751934a92","Type":"ContainerDied","Data":"a95b17fe1e7319a80e3510506126633ec5fda7e4713c1ddeeca6488c4c836110"} Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.698044 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a95b17fe1e7319a80e3510506126633ec5fda7e4713c1ddeeca6488c4c836110" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.697764 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-xsj2v" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.803087 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-88wwf"] Oct 10 16:47:38 crc kubenswrapper[4788]: E1010 16:47:38.803591 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="extract-utilities" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.803611 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="extract-utilities" Oct 10 16:47:38 crc kubenswrapper[4788]: E1010 16:47:38.803642 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="extract-content" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.803649 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="extract-content" Oct 10 16:47:38 crc kubenswrapper[4788]: E1010 16:47:38.803667 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="registry-server" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.803674 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="registry-server" Oct 10 16:47:38 crc kubenswrapper[4788]: E1010 16:47:38.803709 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="171bf601-1ec1-4711-aacd-611751934a92" containerName="bootstrap-openstack-openstack-cell1" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.803717 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="171bf601-1ec1-4711-aacd-611751934a92" containerName="bootstrap-openstack-openstack-cell1" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.803960 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="171bf601-1ec1-4711-aacd-611751934a92" containerName="bootstrap-openstack-openstack-cell1" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.804368 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="921b31bc-b66c-4727-8276-cafb89e8868a" containerName="registry-server" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.806297 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.809393 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.810798 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.812635 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.818531 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-88wwf"] Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.820245 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.942647 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ceph\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.942876 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ssh-key\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.942999 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-inventory\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:38 crc kubenswrapper[4788]: I1010 16:47:38.943245 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwc2j\" (UniqueName: \"kubernetes.io/projected/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-kube-api-access-bwc2j\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.045742 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ssh-key\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.045832 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-inventory\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.045918 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwc2j\" (UniqueName: \"kubernetes.io/projected/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-kube-api-access-bwc2j\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.046034 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ceph\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.052083 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-inventory\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.052540 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ceph\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.058555 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ssh-key\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.064643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwc2j\" (UniqueName: \"kubernetes.io/projected/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-kube-api-access-bwc2j\") pod \"download-cache-openstack-openstack-cell1-88wwf\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.134906 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:47:39 crc kubenswrapper[4788]: I1010 16:47:39.791675 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-88wwf"] Oct 10 16:47:40 crc kubenswrapper[4788]: I1010 16:47:40.742792 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" event={"ID":"d4f46419-8e4b-408e-88aa-55d6bebb8ce9","Type":"ContainerStarted","Data":"19e68d2dcf7038967143e774921923de7d674153eccc009de4b76a759ab99686"} Oct 10 16:47:40 crc kubenswrapper[4788]: I1010 16:47:40.743838 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" event={"ID":"d4f46419-8e4b-408e-88aa-55d6bebb8ce9","Type":"ContainerStarted","Data":"b4e67e3cc6cb6d87ce74ef9c74270c0851f303cd9bfd4bdabdd139571de20b1d"} Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.746899 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" podStartSLOduration=26.537206372 podStartE2EDuration="26.746880156s" podCreationTimestamp="2025-10-10 16:47:38 +0000 UTC" firstStartedPulling="2025-10-10 16:47:39.796104345 +0000 UTC m=+7362.245819893" lastFinishedPulling="2025-10-10 16:47:40.005778139 +0000 UTC m=+7362.455493677" observedRunningTime="2025-10-10 16:47:40.769235804 +0000 UTC m=+7363.218951392" watchObservedRunningTime="2025-10-10 16:48:04.746880156 +0000 UTC m=+7387.196595704" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.751565 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-82tv2"] Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.754016 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.765490 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-82tv2"] Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.859688 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bpm8\" (UniqueName: \"kubernetes.io/projected/50779ee4-ad4c-404f-96de-af0e21f0195a-kube-api-access-4bpm8\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.859753 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-utilities\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.859809 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-catalog-content\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.962640 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bpm8\" (UniqueName: \"kubernetes.io/projected/50779ee4-ad4c-404f-96de-af0e21f0195a-kube-api-access-4bpm8\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.962765 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-utilities\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.962858 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-catalog-content\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.963340 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-utilities\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.963392 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-catalog-content\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:04 crc kubenswrapper[4788]: I1010 16:48:04.988587 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bpm8\" (UniqueName: \"kubernetes.io/projected/50779ee4-ad4c-404f-96de-af0e21f0195a-kube-api-access-4bpm8\") pod \"community-operators-82tv2\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:05 crc kubenswrapper[4788]: I1010 16:48:05.089274 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:05 crc kubenswrapper[4788]: W1010 16:48:05.600502 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50779ee4_ad4c_404f_96de_af0e21f0195a.slice/crio-b4d2a69e2c7b48482489c440a19ce7a6eee2639b55eb7a547ecb29b4046bfd59 WatchSource:0}: Error finding container b4d2a69e2c7b48482489c440a19ce7a6eee2639b55eb7a547ecb29b4046bfd59: Status 404 returned error can't find the container with id b4d2a69e2c7b48482489c440a19ce7a6eee2639b55eb7a547ecb29b4046bfd59 Oct 10 16:48:05 crc kubenswrapper[4788]: I1010 16:48:05.605889 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-82tv2"] Oct 10 16:48:06 crc kubenswrapper[4788]: I1010 16:48:06.048939 4788 generic.go:334] "Generic (PLEG): container finished" podID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerID="4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2" exitCode=0 Oct 10 16:48:06 crc kubenswrapper[4788]: I1010 16:48:06.048985 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82tv2" event={"ID":"50779ee4-ad4c-404f-96de-af0e21f0195a","Type":"ContainerDied","Data":"4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2"} Oct 10 16:48:06 crc kubenswrapper[4788]: I1010 16:48:06.049016 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82tv2" event={"ID":"50779ee4-ad4c-404f-96de-af0e21f0195a","Type":"ContainerStarted","Data":"b4d2a69e2c7b48482489c440a19ce7a6eee2639b55eb7a547ecb29b4046bfd59"} Oct 10 16:48:08 crc kubenswrapper[4788]: I1010 16:48:08.075307 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82tv2" event={"ID":"50779ee4-ad4c-404f-96de-af0e21f0195a","Type":"ContainerStarted","Data":"d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f"} Oct 10 16:48:09 crc kubenswrapper[4788]: I1010 16:48:09.089885 4788 generic.go:334] "Generic (PLEG): container finished" podID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerID="d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f" exitCode=0 Oct 10 16:48:09 crc kubenswrapper[4788]: I1010 16:48:09.090207 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82tv2" event={"ID":"50779ee4-ad4c-404f-96de-af0e21f0195a","Type":"ContainerDied","Data":"d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f"} Oct 10 16:48:10 crc kubenswrapper[4788]: I1010 16:48:10.160734 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82tv2" event={"ID":"50779ee4-ad4c-404f-96de-af0e21f0195a","Type":"ContainerStarted","Data":"5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7"} Oct 10 16:48:10 crc kubenswrapper[4788]: I1010 16:48:10.192412 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-82tv2" podStartSLOduration=2.463974864 podStartE2EDuration="6.192390965s" podCreationTimestamp="2025-10-10 16:48:04 +0000 UTC" firstStartedPulling="2025-10-10 16:48:06.051505233 +0000 UTC m=+7388.501220781" lastFinishedPulling="2025-10-10 16:48:09.779921294 +0000 UTC m=+7392.229636882" observedRunningTime="2025-10-10 16:48:10.180600932 +0000 UTC m=+7392.630316550" watchObservedRunningTime="2025-10-10 16:48:10.192390965 +0000 UTC m=+7392.642106503" Oct 10 16:48:15 crc kubenswrapper[4788]: I1010 16:48:15.090169 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:15 crc kubenswrapper[4788]: I1010 16:48:15.091117 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:15 crc kubenswrapper[4788]: I1010 16:48:15.160460 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:15 crc kubenswrapper[4788]: I1010 16:48:15.304192 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:15 crc kubenswrapper[4788]: I1010 16:48:15.410188 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-82tv2"] Oct 10 16:48:17 crc kubenswrapper[4788]: I1010 16:48:17.252727 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-82tv2" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerName="registry-server" containerID="cri-o://5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7" gracePeriod=2 Oct 10 16:48:17 crc kubenswrapper[4788]: I1010 16:48:17.870075 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:17 crc kubenswrapper[4788]: I1010 16:48:17.915653 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bpm8\" (UniqueName: \"kubernetes.io/projected/50779ee4-ad4c-404f-96de-af0e21f0195a-kube-api-access-4bpm8\") pod \"50779ee4-ad4c-404f-96de-af0e21f0195a\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " Oct 10 16:48:17 crc kubenswrapper[4788]: I1010 16:48:17.915827 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-catalog-content\") pod \"50779ee4-ad4c-404f-96de-af0e21f0195a\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " Oct 10 16:48:17 crc kubenswrapper[4788]: I1010 16:48:17.915880 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-utilities\") pod \"50779ee4-ad4c-404f-96de-af0e21f0195a\" (UID: \"50779ee4-ad4c-404f-96de-af0e21f0195a\") " Oct 10 16:48:17 crc kubenswrapper[4788]: I1010 16:48:17.916975 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-utilities" (OuterVolumeSpecName: "utilities") pod "50779ee4-ad4c-404f-96de-af0e21f0195a" (UID: "50779ee4-ad4c-404f-96de-af0e21f0195a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:48:17 crc kubenswrapper[4788]: I1010 16:48:17.928920 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50779ee4-ad4c-404f-96de-af0e21f0195a-kube-api-access-4bpm8" (OuterVolumeSpecName: "kube-api-access-4bpm8") pod "50779ee4-ad4c-404f-96de-af0e21f0195a" (UID: "50779ee4-ad4c-404f-96de-af0e21f0195a"). InnerVolumeSpecName "kube-api-access-4bpm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:48:17 crc kubenswrapper[4788]: I1010 16:48:17.973556 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50779ee4-ad4c-404f-96de-af0e21f0195a" (UID: "50779ee4-ad4c-404f-96de-af0e21f0195a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.020414 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.020458 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bpm8\" (UniqueName: \"kubernetes.io/projected/50779ee4-ad4c-404f-96de-af0e21f0195a-kube-api-access-4bpm8\") on node \"crc\" DevicePath \"\"" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.020473 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50779ee4-ad4c-404f-96de-af0e21f0195a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.265258 4788 generic.go:334] "Generic (PLEG): container finished" podID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerID="5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7" exitCode=0 Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.265340 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82tv2" event={"ID":"50779ee4-ad4c-404f-96de-af0e21f0195a","Type":"ContainerDied","Data":"5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7"} Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.265387 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82tv2" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.265424 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82tv2" event={"ID":"50779ee4-ad4c-404f-96de-af0e21f0195a","Type":"ContainerDied","Data":"b4d2a69e2c7b48482489c440a19ce7a6eee2639b55eb7a547ecb29b4046bfd59"} Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.265443 4788 scope.go:117] "RemoveContainer" containerID="5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.299469 4788 scope.go:117] "RemoveContainer" containerID="d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.304905 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-82tv2"] Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.314344 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-82tv2"] Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.332611 4788 scope.go:117] "RemoveContainer" containerID="4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.385213 4788 scope.go:117] "RemoveContainer" containerID="5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7" Oct 10 16:48:18 crc kubenswrapper[4788]: E1010 16:48:18.386110 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7\": container with ID starting with 5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7 not found: ID does not exist" containerID="5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.386160 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7"} err="failed to get container status \"5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7\": rpc error: code = NotFound desc = could not find container \"5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7\": container with ID starting with 5f06f3e4fc83c8b4062df28a26c4e03d9b1cb3450d149f7dd2b509fd171605e7 not found: ID does not exist" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.386187 4788 scope.go:117] "RemoveContainer" containerID="d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f" Oct 10 16:48:18 crc kubenswrapper[4788]: E1010 16:48:18.386453 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f\": container with ID starting with d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f not found: ID does not exist" containerID="d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.386470 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f"} err="failed to get container status \"d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f\": rpc error: code = NotFound desc = could not find container \"d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f\": container with ID starting with d08ff47fc49e0a96aacbe5b61a36888c0756eff9e2e7da6d595a9fb6184c9b1f not found: ID does not exist" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.386484 4788 scope.go:117] "RemoveContainer" containerID="4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2" Oct 10 16:48:18 crc kubenswrapper[4788]: E1010 16:48:18.386777 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2\": container with ID starting with 4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2 not found: ID does not exist" containerID="4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2" Oct 10 16:48:18 crc kubenswrapper[4788]: I1010 16:48:18.386801 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2"} err="failed to get container status \"4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2\": rpc error: code = NotFound desc = could not find container \"4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2\": container with ID starting with 4c910679c06ad1f3608a3b9767f8dfac2fc4a8dc2a2e63b16e7bc5b75f30d3f2 not found: ID does not exist" Oct 10 16:48:20 crc kubenswrapper[4788]: I1010 16:48:20.258970 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" path="/var/lib/kubelet/pods/50779ee4-ad4c-404f-96de-af0e21f0195a/volumes" Oct 10 16:48:59 crc kubenswrapper[4788]: I1010 16:48:59.406440 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:48:59 crc kubenswrapper[4788]: I1010 16:48:59.407178 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:49:11 crc kubenswrapper[4788]: I1010 16:49:11.905484 4788 generic.go:334] "Generic (PLEG): container finished" podID="d4f46419-8e4b-408e-88aa-55d6bebb8ce9" containerID="19e68d2dcf7038967143e774921923de7d674153eccc009de4b76a759ab99686" exitCode=0 Oct 10 16:49:11 crc kubenswrapper[4788]: I1010 16:49:11.905535 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" event={"ID":"d4f46419-8e4b-408e-88aa-55d6bebb8ce9","Type":"ContainerDied","Data":"19e68d2dcf7038967143e774921923de7d674153eccc009de4b76a759ab99686"} Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.455203 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.606819 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-inventory\") pod \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.607394 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ceph\") pod \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.607484 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwc2j\" (UniqueName: \"kubernetes.io/projected/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-kube-api-access-bwc2j\") pod \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.607601 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ssh-key\") pod \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\" (UID: \"d4f46419-8e4b-408e-88aa-55d6bebb8ce9\") " Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.613342 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ceph" (OuterVolumeSpecName: "ceph") pod "d4f46419-8e4b-408e-88aa-55d6bebb8ce9" (UID: "d4f46419-8e4b-408e-88aa-55d6bebb8ce9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.614241 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-kube-api-access-bwc2j" (OuterVolumeSpecName: "kube-api-access-bwc2j") pod "d4f46419-8e4b-408e-88aa-55d6bebb8ce9" (UID: "d4f46419-8e4b-408e-88aa-55d6bebb8ce9"). InnerVolumeSpecName "kube-api-access-bwc2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.638895 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d4f46419-8e4b-408e-88aa-55d6bebb8ce9" (UID: "d4f46419-8e4b-408e-88aa-55d6bebb8ce9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.646272 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-inventory" (OuterVolumeSpecName: "inventory") pod "d4f46419-8e4b-408e-88aa-55d6bebb8ce9" (UID: "d4f46419-8e4b-408e-88aa-55d6bebb8ce9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.710641 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.710677 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.710687 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwc2j\" (UniqueName: \"kubernetes.io/projected/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-kube-api-access-bwc2j\") on node \"crc\" DevicePath \"\"" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.710698 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f46419-8e4b-408e-88aa-55d6bebb8ce9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.927720 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" event={"ID":"d4f46419-8e4b-408e-88aa-55d6bebb8ce9","Type":"ContainerDied","Data":"b4e67e3cc6cb6d87ce74ef9c74270c0851f303cd9bfd4bdabdd139571de20b1d"} Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.927765 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4e67e3cc6cb6d87ce74ef9c74270c0851f303cd9bfd4bdabdd139571de20b1d" Oct 10 16:49:13 crc kubenswrapper[4788]: I1010 16:49:13.927794 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-88wwf" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.034746 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-lf2lb"] Oct 10 16:49:14 crc kubenswrapper[4788]: E1010 16:49:14.035245 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerName="registry-server" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.035263 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerName="registry-server" Oct 10 16:49:14 crc kubenswrapper[4788]: E1010 16:49:14.035292 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4f46419-8e4b-408e-88aa-55d6bebb8ce9" containerName="download-cache-openstack-openstack-cell1" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.035300 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4f46419-8e4b-408e-88aa-55d6bebb8ce9" containerName="download-cache-openstack-openstack-cell1" Oct 10 16:49:14 crc kubenswrapper[4788]: E1010 16:49:14.035374 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerName="extract-utilities" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.035383 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerName="extract-utilities" Oct 10 16:49:14 crc kubenswrapper[4788]: E1010 16:49:14.035394 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerName="extract-content" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.035399 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerName="extract-content" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.035644 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="50779ee4-ad4c-404f-96de-af0e21f0195a" containerName="registry-server" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.035670 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4f46419-8e4b-408e-88aa-55d6bebb8ce9" containerName="download-cache-openstack-openstack-cell1" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.036506 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.039598 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.039864 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.039975 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.040086 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.056391 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-lf2lb"] Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.120160 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-inventory\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.120229 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ceph\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.120298 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6gd8\" (UniqueName: \"kubernetes.io/projected/c64cc836-ad93-4dcf-a6f1-29e21606c64c-kube-api-access-q6gd8\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.120552 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ssh-key\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.222910 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ssh-key\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.223134 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-inventory\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.223232 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ceph\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.225719 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6gd8\" (UniqueName: \"kubernetes.io/projected/c64cc836-ad93-4dcf-a6f1-29e21606c64c-kube-api-access-q6gd8\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.227072 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-inventory\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.227727 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ssh-key\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.231623 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ceph\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.243576 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6gd8\" (UniqueName: \"kubernetes.io/projected/c64cc836-ad93-4dcf-a6f1-29e21606c64c-kube-api-access-q6gd8\") pod \"configure-network-openstack-openstack-cell1-lf2lb\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.363124 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.868049 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-lf2lb"] Oct 10 16:49:14 crc kubenswrapper[4788]: I1010 16:49:14.937571 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" event={"ID":"c64cc836-ad93-4dcf-a6f1-29e21606c64c","Type":"ContainerStarted","Data":"71265f67523078c99b0dc09514bc031dcf963ac520f0a9fcd97a2b2aa5729f73"} Oct 10 16:49:15 crc kubenswrapper[4788]: I1010 16:49:15.950200 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" event={"ID":"c64cc836-ad93-4dcf-a6f1-29e21606c64c","Type":"ContainerStarted","Data":"b2ec246afee4c72adb99b4cf24d8df7d45a13aca34c22594ddb8155c83932a07"} Oct 10 16:49:15 crc kubenswrapper[4788]: I1010 16:49:15.974517 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" podStartSLOduration=1.794740026 podStartE2EDuration="1.974281844s" podCreationTimestamp="2025-10-10 16:49:14 +0000 UTC" firstStartedPulling="2025-10-10 16:49:14.875739023 +0000 UTC m=+7457.325454561" lastFinishedPulling="2025-10-10 16:49:15.055280831 +0000 UTC m=+7457.504996379" observedRunningTime="2025-10-10 16:49:15.969498663 +0000 UTC m=+7458.419214211" watchObservedRunningTime="2025-10-10 16:49:15.974281844 +0000 UTC m=+7458.423997392" Oct 10 16:49:29 crc kubenswrapper[4788]: I1010 16:49:29.406026 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:49:29 crc kubenswrapper[4788]: I1010 16:49:29.406927 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:49:59 crc kubenswrapper[4788]: I1010 16:49:59.406892 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:49:59 crc kubenswrapper[4788]: I1010 16:49:59.407942 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:49:59 crc kubenswrapper[4788]: I1010 16:49:59.408040 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:49:59 crc kubenswrapper[4788]: I1010 16:49:59.410233 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e843e596e0077c58474a49e15e5d3b91c305413c7ff87925b3f77a716e6e9ec3"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:49:59 crc kubenswrapper[4788]: I1010 16:49:59.410462 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://e843e596e0077c58474a49e15e5d3b91c305413c7ff87925b3f77a716e6e9ec3" gracePeriod=600 Oct 10 16:50:00 crc kubenswrapper[4788]: I1010 16:50:00.447363 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="e843e596e0077c58474a49e15e5d3b91c305413c7ff87925b3f77a716e6e9ec3" exitCode=0 Oct 10 16:50:00 crc kubenswrapper[4788]: I1010 16:50:00.447429 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"e843e596e0077c58474a49e15e5d3b91c305413c7ff87925b3f77a716e6e9ec3"} Oct 10 16:50:00 crc kubenswrapper[4788]: I1010 16:50:00.448459 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363"} Oct 10 16:50:00 crc kubenswrapper[4788]: I1010 16:50:00.448499 4788 scope.go:117] "RemoveContainer" containerID="a30c7732f1642f4ae81ede054e8de55d6b928ca6901186ae1b7ad4918d51a9fe" Oct 10 16:50:33 crc kubenswrapper[4788]: I1010 16:50:33.885063 4788 generic.go:334] "Generic (PLEG): container finished" podID="c64cc836-ad93-4dcf-a6f1-29e21606c64c" containerID="b2ec246afee4c72adb99b4cf24d8df7d45a13aca34c22594ddb8155c83932a07" exitCode=0 Oct 10 16:50:33 crc kubenswrapper[4788]: I1010 16:50:33.885172 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" event={"ID":"c64cc836-ad93-4dcf-a6f1-29e21606c64c","Type":"ContainerDied","Data":"b2ec246afee4c72adb99b4cf24d8df7d45a13aca34c22594ddb8155c83932a07"} Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.487357 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.648120 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-inventory\") pod \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.648278 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ceph\") pod \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.648315 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ssh-key\") pod \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.648480 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6gd8\" (UniqueName: \"kubernetes.io/projected/c64cc836-ad93-4dcf-a6f1-29e21606c64c-kube-api-access-q6gd8\") pod \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\" (UID: \"c64cc836-ad93-4dcf-a6f1-29e21606c64c\") " Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.659335 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ceph" (OuterVolumeSpecName: "ceph") pod "c64cc836-ad93-4dcf-a6f1-29e21606c64c" (UID: "c64cc836-ad93-4dcf-a6f1-29e21606c64c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.660332 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c64cc836-ad93-4dcf-a6f1-29e21606c64c-kube-api-access-q6gd8" (OuterVolumeSpecName: "kube-api-access-q6gd8") pod "c64cc836-ad93-4dcf-a6f1-29e21606c64c" (UID: "c64cc836-ad93-4dcf-a6f1-29e21606c64c"). InnerVolumeSpecName "kube-api-access-q6gd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.687863 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-inventory" (OuterVolumeSpecName: "inventory") pod "c64cc836-ad93-4dcf-a6f1-29e21606c64c" (UID: "c64cc836-ad93-4dcf-a6f1-29e21606c64c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.700645 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c64cc836-ad93-4dcf-a6f1-29e21606c64c" (UID: "c64cc836-ad93-4dcf-a6f1-29e21606c64c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.751754 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6gd8\" (UniqueName: \"kubernetes.io/projected/c64cc836-ad93-4dcf-a6f1-29e21606c64c-kube-api-access-q6gd8\") on node \"crc\" DevicePath \"\"" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.751801 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.751815 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.751827 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c64cc836-ad93-4dcf-a6f1-29e21606c64c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.912331 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" event={"ID":"c64cc836-ad93-4dcf-a6f1-29e21606c64c","Type":"ContainerDied","Data":"71265f67523078c99b0dc09514bc031dcf963ac520f0a9fcd97a2b2aa5729f73"} Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.912391 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71265f67523078c99b0dc09514bc031dcf963ac520f0a9fcd97a2b2aa5729f73" Oct 10 16:50:35 crc kubenswrapper[4788]: I1010 16:50:35.912498 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-lf2lb" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.027367 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-4dtfv"] Oct 10 16:50:36 crc kubenswrapper[4788]: E1010 16:50:36.028186 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c64cc836-ad93-4dcf-a6f1-29e21606c64c" containerName="configure-network-openstack-openstack-cell1" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.028277 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c64cc836-ad93-4dcf-a6f1-29e21606c64c" containerName="configure-network-openstack-openstack-cell1" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.028557 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c64cc836-ad93-4dcf-a6f1-29e21606c64c" containerName="configure-network-openstack-openstack-cell1" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.029393 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.032558 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.032935 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.033100 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.033262 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.046194 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-4dtfv"] Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.163726 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ceph\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.164118 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ssh-key\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.164358 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-inventory\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.164494 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjlbl\" (UniqueName: \"kubernetes.io/projected/bcdfb941-0c8e-4c5d-8043-19c13616166a-kube-api-access-vjlbl\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.267244 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ceph\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.267316 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ssh-key\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.267406 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-inventory\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.267428 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjlbl\" (UniqueName: \"kubernetes.io/projected/bcdfb941-0c8e-4c5d-8043-19c13616166a-kube-api-access-vjlbl\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.274294 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ssh-key\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.277259 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ceph\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.285547 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-inventory\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.301227 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjlbl\" (UniqueName: \"kubernetes.io/projected/bcdfb941-0c8e-4c5d-8043-19c13616166a-kube-api-access-vjlbl\") pod \"validate-network-openstack-openstack-cell1-4dtfv\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.354998 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.920031 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-4dtfv"] Oct 10 16:50:36 crc kubenswrapper[4788]: I1010 16:50:36.934537 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:50:37 crc kubenswrapper[4788]: I1010 16:50:37.939434 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" event={"ID":"bcdfb941-0c8e-4c5d-8043-19c13616166a","Type":"ContainerStarted","Data":"18f449b81ebceb80d62ee421cbb48429e202a1ed744f859339b0089f8738a65d"} Oct 10 16:50:37 crc kubenswrapper[4788]: I1010 16:50:37.939927 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" event={"ID":"bcdfb941-0c8e-4c5d-8043-19c13616166a","Type":"ContainerStarted","Data":"a394cee25d81b5e44a1b604e00ee9fcf592872df96644c910197f230535b0307"} Oct 10 16:50:37 crc kubenswrapper[4788]: I1010 16:50:37.978477 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" podStartSLOduration=1.7485754 podStartE2EDuration="1.978449809s" podCreationTimestamp="2025-10-10 16:50:36 +0000 UTC" firstStartedPulling="2025-10-10 16:50:36.93427673 +0000 UTC m=+7539.383992278" lastFinishedPulling="2025-10-10 16:50:37.164151129 +0000 UTC m=+7539.613866687" observedRunningTime="2025-10-10 16:50:37.961550204 +0000 UTC m=+7540.411265762" watchObservedRunningTime="2025-10-10 16:50:37.978449809 +0000 UTC m=+7540.428165367" Oct 10 16:50:42 crc kubenswrapper[4788]: I1010 16:50:42.995380 4788 generic.go:334] "Generic (PLEG): container finished" podID="bcdfb941-0c8e-4c5d-8043-19c13616166a" containerID="18f449b81ebceb80d62ee421cbb48429e202a1ed744f859339b0089f8738a65d" exitCode=0 Oct 10 16:50:42 crc kubenswrapper[4788]: I1010 16:50:42.995459 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" event={"ID":"bcdfb941-0c8e-4c5d-8043-19c13616166a","Type":"ContainerDied","Data":"18f449b81ebceb80d62ee421cbb48429e202a1ed744f859339b0089f8738a65d"} Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.441728 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.563430 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ssh-key\") pod \"bcdfb941-0c8e-4c5d-8043-19c13616166a\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.563504 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-inventory\") pod \"bcdfb941-0c8e-4c5d-8043-19c13616166a\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.563567 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjlbl\" (UniqueName: \"kubernetes.io/projected/bcdfb941-0c8e-4c5d-8043-19c13616166a-kube-api-access-vjlbl\") pod \"bcdfb941-0c8e-4c5d-8043-19c13616166a\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.563715 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ceph\") pod \"bcdfb941-0c8e-4c5d-8043-19c13616166a\" (UID: \"bcdfb941-0c8e-4c5d-8043-19c13616166a\") " Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.568677 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcdfb941-0c8e-4c5d-8043-19c13616166a-kube-api-access-vjlbl" (OuterVolumeSpecName: "kube-api-access-vjlbl") pod "bcdfb941-0c8e-4c5d-8043-19c13616166a" (UID: "bcdfb941-0c8e-4c5d-8043-19c13616166a"). InnerVolumeSpecName "kube-api-access-vjlbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.568719 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ceph" (OuterVolumeSpecName: "ceph") pod "bcdfb941-0c8e-4c5d-8043-19c13616166a" (UID: "bcdfb941-0c8e-4c5d-8043-19c13616166a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.591252 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-inventory" (OuterVolumeSpecName: "inventory") pod "bcdfb941-0c8e-4c5d-8043-19c13616166a" (UID: "bcdfb941-0c8e-4c5d-8043-19c13616166a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.592682 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bcdfb941-0c8e-4c5d-8043-19c13616166a" (UID: "bcdfb941-0c8e-4c5d-8043-19c13616166a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.666702 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.666913 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.666982 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjlbl\" (UniqueName: \"kubernetes.io/projected/bcdfb941-0c8e-4c5d-8043-19c13616166a-kube-api-access-vjlbl\") on node \"crc\" DevicePath \"\"" Oct 10 16:50:44 crc kubenswrapper[4788]: I1010 16:50:44.667060 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bcdfb941-0c8e-4c5d-8043-19c13616166a-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.017472 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" event={"ID":"bcdfb941-0c8e-4c5d-8043-19c13616166a","Type":"ContainerDied","Data":"a394cee25d81b5e44a1b604e00ee9fcf592872df96644c910197f230535b0307"} Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.017537 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a394cee25d81b5e44a1b604e00ee9fcf592872df96644c910197f230535b0307" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.017552 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-4dtfv" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.097963 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-84z6g"] Oct 10 16:50:45 crc kubenswrapper[4788]: E1010 16:50:45.098383 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcdfb941-0c8e-4c5d-8043-19c13616166a" containerName="validate-network-openstack-openstack-cell1" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.098399 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcdfb941-0c8e-4c5d-8043-19c13616166a" containerName="validate-network-openstack-openstack-cell1" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.098673 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcdfb941-0c8e-4c5d-8043-19c13616166a" containerName="validate-network-openstack-openstack-cell1" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.099584 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.102800 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.104943 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.105054 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.105936 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.114417 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-84z6g"] Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.178275 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ssh-key\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.178352 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ceph\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.178420 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-inventory\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.178563 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8lrh\" (UniqueName: \"kubernetes.io/projected/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-kube-api-access-g8lrh\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.280727 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8lrh\" (UniqueName: \"kubernetes.io/projected/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-kube-api-access-g8lrh\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.280937 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ssh-key\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.280975 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ceph\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.281022 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-inventory\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.286797 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ssh-key\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.287746 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-inventory\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.290803 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ceph\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.297334 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8lrh\" (UniqueName: \"kubernetes.io/projected/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-kube-api-access-g8lrh\") pod \"install-os-openstack-openstack-cell1-84z6g\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.424490 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:50:45 crc kubenswrapper[4788]: I1010 16:50:45.987854 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-84z6g"] Oct 10 16:50:46 crc kubenswrapper[4788]: I1010 16:50:46.031128 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-84z6g" event={"ID":"2793e9af-a0b7-4dc7-91fc-44ea0f36366a","Type":"ContainerStarted","Data":"b20b01de1904b4cca72015fbe55ffa15c9ae8de1cd182143754a58409839bfe0"} Oct 10 16:50:47 crc kubenswrapper[4788]: I1010 16:50:47.049083 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-84z6g" event={"ID":"2793e9af-a0b7-4dc7-91fc-44ea0f36366a","Type":"ContainerStarted","Data":"1c0669d04c938cb01eb29a987c62991ca08a64cfbb4c33ddf0450e6a4f70fb86"} Oct 10 16:50:47 crc kubenswrapper[4788]: I1010 16:50:47.070495 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-84z6g" podStartSLOduration=1.886738859 podStartE2EDuration="2.070473682s" podCreationTimestamp="2025-10-10 16:50:45 +0000 UTC" firstStartedPulling="2025-10-10 16:50:45.998626684 +0000 UTC m=+7548.448342252" lastFinishedPulling="2025-10-10 16:50:46.182361527 +0000 UTC m=+7548.632077075" observedRunningTime="2025-10-10 16:50:47.064993572 +0000 UTC m=+7549.514709120" watchObservedRunningTime="2025-10-10 16:50:47.070473682 +0000 UTC m=+7549.520189240" Oct 10 16:51:31 crc kubenswrapper[4788]: I1010 16:51:31.512925 4788 generic.go:334] "Generic (PLEG): container finished" podID="2793e9af-a0b7-4dc7-91fc-44ea0f36366a" containerID="1c0669d04c938cb01eb29a987c62991ca08a64cfbb4c33ddf0450e6a4f70fb86" exitCode=0 Oct 10 16:51:31 crc kubenswrapper[4788]: I1010 16:51:31.513007 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-84z6g" event={"ID":"2793e9af-a0b7-4dc7-91fc-44ea0f36366a","Type":"ContainerDied","Data":"1c0669d04c938cb01eb29a987c62991ca08a64cfbb4c33ddf0450e6a4f70fb86"} Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.031600 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.096392 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ssh-key\") pod \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.096955 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-inventory\") pod \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.097043 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8lrh\" (UniqueName: \"kubernetes.io/projected/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-kube-api-access-g8lrh\") pod \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.097304 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ceph\") pod \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\" (UID: \"2793e9af-a0b7-4dc7-91fc-44ea0f36366a\") " Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.103625 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ceph" (OuterVolumeSpecName: "ceph") pod "2793e9af-a0b7-4dc7-91fc-44ea0f36366a" (UID: "2793e9af-a0b7-4dc7-91fc-44ea0f36366a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.105492 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-kube-api-access-g8lrh" (OuterVolumeSpecName: "kube-api-access-g8lrh") pod "2793e9af-a0b7-4dc7-91fc-44ea0f36366a" (UID: "2793e9af-a0b7-4dc7-91fc-44ea0f36366a"). InnerVolumeSpecName "kube-api-access-g8lrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.127444 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2793e9af-a0b7-4dc7-91fc-44ea0f36366a" (UID: "2793e9af-a0b7-4dc7-91fc-44ea0f36366a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.139814 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-inventory" (OuterVolumeSpecName: "inventory") pod "2793e9af-a0b7-4dc7-91fc-44ea0f36366a" (UID: "2793e9af-a0b7-4dc7-91fc-44ea0f36366a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.201327 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8lrh\" (UniqueName: \"kubernetes.io/projected/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-kube-api-access-g8lrh\") on node \"crc\" DevicePath \"\"" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.201434 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.201459 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.201485 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2793e9af-a0b7-4dc7-91fc-44ea0f36366a-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.560668 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-84z6g" event={"ID":"2793e9af-a0b7-4dc7-91fc-44ea0f36366a","Type":"ContainerDied","Data":"b20b01de1904b4cca72015fbe55ffa15c9ae8de1cd182143754a58409839bfe0"} Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.561054 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b20b01de1904b4cca72015fbe55ffa15c9ae8de1cd182143754a58409839bfe0" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.560749 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-84z6g" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.622901 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-n5thk"] Oct 10 16:51:33 crc kubenswrapper[4788]: E1010 16:51:33.623441 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2793e9af-a0b7-4dc7-91fc-44ea0f36366a" containerName="install-os-openstack-openstack-cell1" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.623459 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2793e9af-a0b7-4dc7-91fc-44ea0f36366a" containerName="install-os-openstack-openstack-cell1" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.623725 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="2793e9af-a0b7-4dc7-91fc-44ea0f36366a" containerName="install-os-openstack-openstack-cell1" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.624757 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.627786 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.628061 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.628273 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.628488 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.642616 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-n5thk"] Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.711669 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-inventory\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.711758 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ceph\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.711799 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ssh-key\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.712107 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq8bc\" (UniqueName: \"kubernetes.io/projected/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-kube-api-access-qq8bc\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.814847 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq8bc\" (UniqueName: \"kubernetes.io/projected/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-kube-api-access-qq8bc\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.815266 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-inventory\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.816827 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ceph\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.817054 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ssh-key\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.819385 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-inventory\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.819822 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ssh-key\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.820397 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ceph\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.831750 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq8bc\" (UniqueName: \"kubernetes.io/projected/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-kube-api-access-qq8bc\") pod \"configure-os-openstack-openstack-cell1-n5thk\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:33 crc kubenswrapper[4788]: I1010 16:51:33.949464 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:51:34 crc kubenswrapper[4788]: I1010 16:51:34.482570 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-n5thk"] Oct 10 16:51:34 crc kubenswrapper[4788]: I1010 16:51:34.570285 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" event={"ID":"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6","Type":"ContainerStarted","Data":"6450e92e4c65c24b2b169389646ed26fb3a840b7683fee2e77ee615678404235"} Oct 10 16:51:35 crc kubenswrapper[4788]: I1010 16:51:35.583214 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" event={"ID":"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6","Type":"ContainerStarted","Data":"01fd1cd5ea163c6126e024f227b160f8d44c2c999faa323baffdce23c888b5a9"} Oct 10 16:51:35 crc kubenswrapper[4788]: I1010 16:51:35.616461 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" podStartSLOduration=2.462494502 podStartE2EDuration="2.616439727s" podCreationTimestamp="2025-10-10 16:51:33 +0000 UTC" firstStartedPulling="2025-10-10 16:51:34.492125189 +0000 UTC m=+7596.941840737" lastFinishedPulling="2025-10-10 16:51:34.646070414 +0000 UTC m=+7597.095785962" observedRunningTime="2025-10-10 16:51:35.607298457 +0000 UTC m=+7598.057014065" watchObservedRunningTime="2025-10-10 16:51:35.616439727 +0000 UTC m=+7598.066155275" Oct 10 16:51:59 crc kubenswrapper[4788]: I1010 16:51:59.406730 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:51:59 crc kubenswrapper[4788]: I1010 16:51:59.408554 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:52:21 crc kubenswrapper[4788]: I1010 16:52:21.118766 4788 generic.go:334] "Generic (PLEG): container finished" podID="d6bf08dd-da00-4e04-856f-8b6bff2eb1e6" containerID="01fd1cd5ea163c6126e024f227b160f8d44c2c999faa323baffdce23c888b5a9" exitCode=0 Oct 10 16:52:21 crc kubenswrapper[4788]: I1010 16:52:21.119372 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" event={"ID":"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6","Type":"ContainerDied","Data":"01fd1cd5ea163c6126e024f227b160f8d44c2c999faa323baffdce23c888b5a9"} Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.641080 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.744621 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq8bc\" (UniqueName: \"kubernetes.io/projected/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-kube-api-access-qq8bc\") pod \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.744796 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ceph\") pod \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.744832 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-inventory\") pod \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.744867 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ssh-key\") pod \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\" (UID: \"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6\") " Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.752350 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ceph" (OuterVolumeSpecName: "ceph") pod "d6bf08dd-da00-4e04-856f-8b6bff2eb1e6" (UID: "d6bf08dd-da00-4e04-856f-8b6bff2eb1e6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.752423 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-kube-api-access-qq8bc" (OuterVolumeSpecName: "kube-api-access-qq8bc") pod "d6bf08dd-da00-4e04-856f-8b6bff2eb1e6" (UID: "d6bf08dd-da00-4e04-856f-8b6bff2eb1e6"). InnerVolumeSpecName "kube-api-access-qq8bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.774319 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-inventory" (OuterVolumeSpecName: "inventory") pod "d6bf08dd-da00-4e04-856f-8b6bff2eb1e6" (UID: "d6bf08dd-da00-4e04-856f-8b6bff2eb1e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.776105 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6bf08dd-da00-4e04-856f-8b6bff2eb1e6" (UID: "d6bf08dd-da00-4e04-856f-8b6bff2eb1e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.847752 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.847780 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.847789 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:22 crc kubenswrapper[4788]: I1010 16:52:22.847799 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq8bc\" (UniqueName: \"kubernetes.io/projected/d6bf08dd-da00-4e04-856f-8b6bff2eb1e6-kube-api-access-qq8bc\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.143325 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" event={"ID":"d6bf08dd-da00-4e04-856f-8b6bff2eb1e6","Type":"ContainerDied","Data":"6450e92e4c65c24b2b169389646ed26fb3a840b7683fee2e77ee615678404235"} Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.143363 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6450e92e4c65c24b2b169389646ed26fb3a840b7683fee2e77ee615678404235" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.143393 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-n5thk" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.245929 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-rz6fx"] Oct 10 16:52:23 crc kubenswrapper[4788]: E1010 16:52:23.246542 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6bf08dd-da00-4e04-856f-8b6bff2eb1e6" containerName="configure-os-openstack-openstack-cell1" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.246566 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6bf08dd-da00-4e04-856f-8b6bff2eb1e6" containerName="configure-os-openstack-openstack-cell1" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.246865 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6bf08dd-da00-4e04-856f-8b6bff2eb1e6" containerName="configure-os-openstack-openstack-cell1" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.247975 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.251108 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.251599 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.251648 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.251779 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.253177 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-rz6fx"] Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.357731 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.358052 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-inventory-0\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.358252 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgs8l\" (UniqueName: \"kubernetes.io/projected/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-kube-api-access-zgs8l\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.358352 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ceph\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: E1010 16:52:23.364724 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6bf08dd_da00_4e04_856f_8b6bff2eb1e6.slice/crio-6450e92e4c65c24b2b169389646ed26fb3a840b7683fee2e77ee615678404235\": RecentStats: unable to find data in memory cache]" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.460851 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgs8l\" (UniqueName: \"kubernetes.io/projected/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-kube-api-access-zgs8l\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.461183 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ceph\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.461477 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.461540 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-inventory-0\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.465003 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ceph\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.465106 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-inventory-0\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.474064 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.477175 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgs8l\" (UniqueName: \"kubernetes.io/projected/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-kube-api-access-zgs8l\") pod \"ssh-known-hosts-openstack-rz6fx\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:23 crc kubenswrapper[4788]: I1010 16:52:23.582301 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:24 crc kubenswrapper[4788]: I1010 16:52:24.326862 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-rz6fx"] Oct 10 16:52:25 crc kubenswrapper[4788]: I1010 16:52:25.186220 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rz6fx" event={"ID":"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d","Type":"ContainerStarted","Data":"6c2c333f7bca46da2b6f4152426f1025a38a23a22597bfe6a5ba1eb771e91e70"} Oct 10 16:52:25 crc kubenswrapper[4788]: I1010 16:52:25.186266 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rz6fx" event={"ID":"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d","Type":"ContainerStarted","Data":"a63353fc3367e230e73c4e8afe770639fdb6d155187371f1cd60f688567f2990"} Oct 10 16:52:25 crc kubenswrapper[4788]: I1010 16:52:25.228476 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-rz6fx" podStartSLOduration=2.044436652 podStartE2EDuration="2.228452682s" podCreationTimestamp="2025-10-10 16:52:23 +0000 UTC" firstStartedPulling="2025-10-10 16:52:24.342214177 +0000 UTC m=+7646.791929725" lastFinishedPulling="2025-10-10 16:52:24.526230197 +0000 UTC m=+7646.975945755" observedRunningTime="2025-10-10 16:52:25.211462465 +0000 UTC m=+7647.661178023" watchObservedRunningTime="2025-10-10 16:52:25.228452682 +0000 UTC m=+7647.678168230" Oct 10 16:52:29 crc kubenswrapper[4788]: I1010 16:52:29.406740 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:52:29 crc kubenswrapper[4788]: I1010 16:52:29.407350 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:52:34 crc kubenswrapper[4788]: I1010 16:52:34.278928 4788 generic.go:334] "Generic (PLEG): container finished" podID="1eb686cf-9ffe-4b1d-ad32-3800206d3c7d" containerID="6c2c333f7bca46da2b6f4152426f1025a38a23a22597bfe6a5ba1eb771e91e70" exitCode=0 Oct 10 16:52:34 crc kubenswrapper[4788]: I1010 16:52:34.279038 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rz6fx" event={"ID":"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d","Type":"ContainerDied","Data":"6c2c333f7bca46da2b6f4152426f1025a38a23a22597bfe6a5ba1eb771e91e70"} Oct 10 16:52:35 crc kubenswrapper[4788]: I1010 16:52:35.843527 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:35 crc kubenswrapper[4788]: I1010 16:52:35.955646 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-inventory-0\") pod \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " Oct 10 16:52:35 crc kubenswrapper[4788]: I1010 16:52:35.955726 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgs8l\" (UniqueName: \"kubernetes.io/projected/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-kube-api-access-zgs8l\") pod \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " Oct 10 16:52:35 crc kubenswrapper[4788]: I1010 16:52:35.955824 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ceph\") pod \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " Oct 10 16:52:35 crc kubenswrapper[4788]: I1010 16:52:35.955931 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ssh-key-openstack-cell1\") pod \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\" (UID: \"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d\") " Oct 10 16:52:35 crc kubenswrapper[4788]: I1010 16:52:35.962390 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ceph" (OuterVolumeSpecName: "ceph") pod "1eb686cf-9ffe-4b1d-ad32-3800206d3c7d" (UID: "1eb686cf-9ffe-4b1d-ad32-3800206d3c7d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:35 crc kubenswrapper[4788]: I1010 16:52:35.963402 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-kube-api-access-zgs8l" (OuterVolumeSpecName: "kube-api-access-zgs8l") pod "1eb686cf-9ffe-4b1d-ad32-3800206d3c7d" (UID: "1eb686cf-9ffe-4b1d-ad32-3800206d3c7d"). InnerVolumeSpecName "kube-api-access-zgs8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.005313 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "1eb686cf-9ffe-4b1d-ad32-3800206d3c7d" (UID: "1eb686cf-9ffe-4b1d-ad32-3800206d3c7d"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.029316 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "1eb686cf-9ffe-4b1d-ad32-3800206d3c7d" (UID: "1eb686cf-9ffe-4b1d-ad32-3800206d3c7d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.060212 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.060427 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.060582 4788 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.060697 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgs8l\" (UniqueName: \"kubernetes.io/projected/1eb686cf-9ffe-4b1d-ad32-3800206d3c7d-kube-api-access-zgs8l\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.308575 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rz6fx" event={"ID":"1eb686cf-9ffe-4b1d-ad32-3800206d3c7d","Type":"ContainerDied","Data":"a63353fc3367e230e73c4e8afe770639fdb6d155187371f1cd60f688567f2990"} Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.308830 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a63353fc3367e230e73c4e8afe770639fdb6d155187371f1cd60f688567f2990" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.308667 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rz6fx" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.369659 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-dnvvt"] Oct 10 16:52:36 crc kubenswrapper[4788]: E1010 16:52:36.370323 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb686cf-9ffe-4b1d-ad32-3800206d3c7d" containerName="ssh-known-hosts-openstack" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.371069 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb686cf-9ffe-4b1d-ad32-3800206d3c7d" containerName="ssh-known-hosts-openstack" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.371409 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb686cf-9ffe-4b1d-ad32-3800206d3c7d" containerName="ssh-known-hosts-openstack" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.372236 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.375796 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.375825 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.375832 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.377390 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.397477 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-dnvvt"] Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.470005 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l5fn\" (UniqueName: \"kubernetes.io/projected/480d7aad-e180-4e43-8179-b6e1d63dc616-kube-api-access-6l5fn\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.470397 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-inventory\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.470753 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ssh-key\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.470859 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ceph\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.572868 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-inventory\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.573227 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ssh-key\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.573422 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ceph\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.573665 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l5fn\" (UniqueName: \"kubernetes.io/projected/480d7aad-e180-4e43-8179-b6e1d63dc616-kube-api-access-6l5fn\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.578340 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-inventory\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.578404 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ssh-key\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.578463 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ceph\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.592813 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l5fn\" (UniqueName: \"kubernetes.io/projected/480d7aad-e180-4e43-8179-b6e1d63dc616-kube-api-access-6l5fn\") pod \"run-os-openstack-openstack-cell1-dnvvt\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:36 crc kubenswrapper[4788]: I1010 16:52:36.694701 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:37 crc kubenswrapper[4788]: W1010 16:52:37.280004 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod480d7aad_e180_4e43_8179_b6e1d63dc616.slice/crio-7566d2aa5380ec3435725440e93808b3a9e05dc95b3fc47cedbb1f7eb75b2197 WatchSource:0}: Error finding container 7566d2aa5380ec3435725440e93808b3a9e05dc95b3fc47cedbb1f7eb75b2197: Status 404 returned error can't find the container with id 7566d2aa5380ec3435725440e93808b3a9e05dc95b3fc47cedbb1f7eb75b2197 Oct 10 16:52:37 crc kubenswrapper[4788]: I1010 16:52:37.283939 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-dnvvt"] Oct 10 16:52:37 crc kubenswrapper[4788]: I1010 16:52:37.325642 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" event={"ID":"480d7aad-e180-4e43-8179-b6e1d63dc616","Type":"ContainerStarted","Data":"7566d2aa5380ec3435725440e93808b3a9e05dc95b3fc47cedbb1f7eb75b2197"} Oct 10 16:52:38 crc kubenswrapper[4788]: I1010 16:52:38.344814 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" event={"ID":"480d7aad-e180-4e43-8179-b6e1d63dc616","Type":"ContainerStarted","Data":"a3ec08e7dcc8238feda65e0f660c57ab73fbcbee52b0e74a2b76cf954801dae3"} Oct 10 16:52:38 crc kubenswrapper[4788]: I1010 16:52:38.398816 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" podStartSLOduration=2.235685364 podStartE2EDuration="2.39879141s" podCreationTimestamp="2025-10-10 16:52:36 +0000 UTC" firstStartedPulling="2025-10-10 16:52:37.282558704 +0000 UTC m=+7659.732274272" lastFinishedPulling="2025-10-10 16:52:37.44566477 +0000 UTC m=+7659.895380318" observedRunningTime="2025-10-10 16:52:38.370451423 +0000 UTC m=+7660.820167011" watchObservedRunningTime="2025-10-10 16:52:38.39879141 +0000 UTC m=+7660.848506958" Oct 10 16:52:46 crc kubenswrapper[4788]: I1010 16:52:46.431629 4788 generic.go:334] "Generic (PLEG): container finished" podID="480d7aad-e180-4e43-8179-b6e1d63dc616" containerID="a3ec08e7dcc8238feda65e0f660c57ab73fbcbee52b0e74a2b76cf954801dae3" exitCode=0 Oct 10 16:52:46 crc kubenswrapper[4788]: I1010 16:52:46.431715 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" event={"ID":"480d7aad-e180-4e43-8179-b6e1d63dc616","Type":"ContainerDied","Data":"a3ec08e7dcc8238feda65e0f660c57ab73fbcbee52b0e74a2b76cf954801dae3"} Oct 10 16:52:47 crc kubenswrapper[4788]: I1010 16:52:47.944232 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.039707 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-inventory\") pod \"480d7aad-e180-4e43-8179-b6e1d63dc616\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.039815 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l5fn\" (UniqueName: \"kubernetes.io/projected/480d7aad-e180-4e43-8179-b6e1d63dc616-kube-api-access-6l5fn\") pod \"480d7aad-e180-4e43-8179-b6e1d63dc616\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.039933 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ssh-key\") pod \"480d7aad-e180-4e43-8179-b6e1d63dc616\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.039991 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ceph\") pod \"480d7aad-e180-4e43-8179-b6e1d63dc616\" (UID: \"480d7aad-e180-4e43-8179-b6e1d63dc616\") " Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.047041 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ceph" (OuterVolumeSpecName: "ceph") pod "480d7aad-e180-4e43-8179-b6e1d63dc616" (UID: "480d7aad-e180-4e43-8179-b6e1d63dc616"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.047666 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480d7aad-e180-4e43-8179-b6e1d63dc616-kube-api-access-6l5fn" (OuterVolumeSpecName: "kube-api-access-6l5fn") pod "480d7aad-e180-4e43-8179-b6e1d63dc616" (UID: "480d7aad-e180-4e43-8179-b6e1d63dc616"). InnerVolumeSpecName "kube-api-access-6l5fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.083528 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "480d7aad-e180-4e43-8179-b6e1d63dc616" (UID: "480d7aad-e180-4e43-8179-b6e1d63dc616"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.097621 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-inventory" (OuterVolumeSpecName: "inventory") pod "480d7aad-e180-4e43-8179-b6e1d63dc616" (UID: "480d7aad-e180-4e43-8179-b6e1d63dc616"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.142168 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l5fn\" (UniqueName: \"kubernetes.io/projected/480d7aad-e180-4e43-8179-b6e1d63dc616-kube-api-access-6l5fn\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.142208 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.142220 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.142231 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480d7aad-e180-4e43-8179-b6e1d63dc616-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.458982 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" event={"ID":"480d7aad-e180-4e43-8179-b6e1d63dc616","Type":"ContainerDied","Data":"7566d2aa5380ec3435725440e93808b3a9e05dc95b3fc47cedbb1f7eb75b2197"} Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.459587 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7566d2aa5380ec3435725440e93808b3a9e05dc95b3fc47cedbb1f7eb75b2197" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.459067 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-dnvvt" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.534730 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-d5bk5"] Oct 10 16:52:48 crc kubenswrapper[4788]: E1010 16:52:48.535266 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480d7aad-e180-4e43-8179-b6e1d63dc616" containerName="run-os-openstack-openstack-cell1" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.535285 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="480d7aad-e180-4e43-8179-b6e1d63dc616" containerName="run-os-openstack-openstack-cell1" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.535552 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="480d7aad-e180-4e43-8179-b6e1d63dc616" containerName="run-os-openstack-openstack-cell1" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.536528 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.539565 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.539703 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.539784 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.539930 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.549908 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-d5bk5"] Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.659574 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-inventory\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.659842 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k59vr\" (UniqueName: \"kubernetes.io/projected/38fe23d3-2d05-463f-9bea-67a8fe0cb377-kube-api-access-k59vr\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.660223 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ceph\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.660318 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.763433 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k59vr\" (UniqueName: \"kubernetes.io/projected/38fe23d3-2d05-463f-9bea-67a8fe0cb377-kube-api-access-k59vr\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.763555 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ceph\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.763603 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.763684 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-inventory\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.768812 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.770643 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ceph\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.772214 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-inventory\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.787642 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k59vr\" (UniqueName: \"kubernetes.io/projected/38fe23d3-2d05-463f-9bea-67a8fe0cb377-kube-api-access-k59vr\") pod \"reboot-os-openstack-openstack-cell1-d5bk5\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:48 crc kubenswrapper[4788]: I1010 16:52:48.867787 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:52:50 crc kubenswrapper[4788]: I1010 16:52:50.422892 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-d5bk5"] Oct 10 16:52:50 crc kubenswrapper[4788]: I1010 16:52:50.489318 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" event={"ID":"38fe23d3-2d05-463f-9bea-67a8fe0cb377","Type":"ContainerStarted","Data":"43c625dde25a7832a70e57433a72a1124497f004debda8fbb0eaa468ddece17a"} Oct 10 16:52:51 crc kubenswrapper[4788]: I1010 16:52:51.502992 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" event={"ID":"38fe23d3-2d05-463f-9bea-67a8fe0cb377","Type":"ContainerStarted","Data":"abda23024f151df5df71fccb38f3494f40e6f8aad59ca14529d9ba9952aff6c6"} Oct 10 16:52:51 crc kubenswrapper[4788]: I1010 16:52:51.531977 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" podStartSLOduration=3.345596063 podStartE2EDuration="3.531943948s" podCreationTimestamp="2025-10-10 16:52:48 +0000 UTC" firstStartedPulling="2025-10-10 16:52:50.428660137 +0000 UTC m=+7672.878375705" lastFinishedPulling="2025-10-10 16:52:50.615008042 +0000 UTC m=+7673.064723590" observedRunningTime="2025-10-10 16:52:51.523324552 +0000 UTC m=+7673.973040100" watchObservedRunningTime="2025-10-10 16:52:51.531943948 +0000 UTC m=+7673.981659536" Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.406447 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.407188 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.407236 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.407902 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.407962 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" gracePeriod=600 Oct 10 16:52:59 crc kubenswrapper[4788]: E1010 16:52:59.540493 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.592372 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363"} Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.592462 4788 scope.go:117] "RemoveContainer" containerID="e843e596e0077c58474a49e15e5d3b91c305413c7ff87925b3f77a716e6e9ec3" Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.593285 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:52:59 crc kubenswrapper[4788]: E1010 16:52:59.593892 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:52:59 crc kubenswrapper[4788]: I1010 16:52:59.591722 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" exitCode=0 Oct 10 16:53:07 crc kubenswrapper[4788]: I1010 16:53:07.700685 4788 generic.go:334] "Generic (PLEG): container finished" podID="38fe23d3-2d05-463f-9bea-67a8fe0cb377" containerID="abda23024f151df5df71fccb38f3494f40e6f8aad59ca14529d9ba9952aff6c6" exitCode=0 Oct 10 16:53:07 crc kubenswrapper[4788]: I1010 16:53:07.700842 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" event={"ID":"38fe23d3-2d05-463f-9bea-67a8fe0cb377","Type":"ContainerDied","Data":"abda23024f151df5df71fccb38f3494f40e6f8aad59ca14529d9ba9952aff6c6"} Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.158273 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.211893 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k59vr\" (UniqueName: \"kubernetes.io/projected/38fe23d3-2d05-463f-9bea-67a8fe0cb377-kube-api-access-k59vr\") pod \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.220517 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38fe23d3-2d05-463f-9bea-67a8fe0cb377-kube-api-access-k59vr" (OuterVolumeSpecName: "kube-api-access-k59vr") pod "38fe23d3-2d05-463f-9bea-67a8fe0cb377" (UID: "38fe23d3-2d05-463f-9bea-67a8fe0cb377"). InnerVolumeSpecName "kube-api-access-k59vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.317268 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ssh-key\") pod \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.317592 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-inventory\") pod \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.317647 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ceph\") pod \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\" (UID: \"38fe23d3-2d05-463f-9bea-67a8fe0cb377\") " Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.318892 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k59vr\" (UniqueName: \"kubernetes.io/projected/38fe23d3-2d05-463f-9bea-67a8fe0cb377-kube-api-access-k59vr\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.324497 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ceph" (OuterVolumeSpecName: "ceph") pod "38fe23d3-2d05-463f-9bea-67a8fe0cb377" (UID: "38fe23d3-2d05-463f-9bea-67a8fe0cb377"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.376495 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "38fe23d3-2d05-463f-9bea-67a8fe0cb377" (UID: "38fe23d3-2d05-463f-9bea-67a8fe0cb377"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.392942 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-inventory" (OuterVolumeSpecName: "inventory") pod "38fe23d3-2d05-463f-9bea-67a8fe0cb377" (UID: "38fe23d3-2d05-463f-9bea-67a8fe0cb377"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.421692 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.421833 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.421898 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/38fe23d3-2d05-463f-9bea-67a8fe0cb377-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.726429 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" event={"ID":"38fe23d3-2d05-463f-9bea-67a8fe0cb377","Type":"ContainerDied","Data":"43c625dde25a7832a70e57433a72a1124497f004debda8fbb0eaa468ddece17a"} Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.726481 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-d5bk5" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.726478 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43c625dde25a7832a70e57433a72a1124497f004debda8fbb0eaa468ddece17a" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.872284 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-97dq9"] Oct 10 16:53:09 crc kubenswrapper[4788]: E1010 16:53:09.872864 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38fe23d3-2d05-463f-9bea-67a8fe0cb377" containerName="reboot-os-openstack-openstack-cell1" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.872885 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="38fe23d3-2d05-463f-9bea-67a8fe0cb377" containerName="reboot-os-openstack-openstack-cell1" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.873287 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="38fe23d3-2d05-463f-9bea-67a8fe0cb377" containerName="reboot-os-openstack-openstack-cell1" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.874352 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.879283 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.879389 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.882558 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.893783 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:53:09 crc kubenswrapper[4788]: I1010 16:53:09.897584 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-97dq9"] Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.035469 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.035615 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.035661 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.035697 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ceph\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.035830 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.035892 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.035972 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.036017 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ssh-key\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.036115 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.036190 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.036292 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-inventory\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.036346 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65ngn\" (UniqueName: \"kubernetes.io/projected/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-kube-api-access-65ngn\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139275 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139585 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139627 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139650 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ssh-key\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139706 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139740 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139826 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-inventory\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139872 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65ngn\" (UniqueName: \"kubernetes.io/projected/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-kube-api-access-65ngn\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139920 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.139972 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.140001 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.140028 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ceph\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.144865 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ceph\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.144887 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.146699 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-inventory\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.146992 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.148246 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.148374 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ssh-key\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.150351 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.150628 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.150964 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.155390 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.161004 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.165214 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65ngn\" (UniqueName: \"kubernetes.io/projected/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-kube-api-access-65ngn\") pod \"install-certs-openstack-openstack-cell1-97dq9\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.191562 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:10 crc kubenswrapper[4788]: I1010 16:53:10.848452 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-97dq9"] Oct 10 16:53:11 crc kubenswrapper[4788]: I1010 16:53:11.237060 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:53:11 crc kubenswrapper[4788]: E1010 16:53:11.237838 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:53:11 crc kubenswrapper[4788]: I1010 16:53:11.756887 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" event={"ID":"6eff8747-d0bb-49d9-b45d-95ed5adf39a0","Type":"ContainerStarted","Data":"562d60a8b2233da9c94b2d639317a068489aa8b87a13854ad59eb0d802538d3d"} Oct 10 16:53:11 crc kubenswrapper[4788]: I1010 16:53:11.757262 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" event={"ID":"6eff8747-d0bb-49d9-b45d-95ed5adf39a0","Type":"ContainerStarted","Data":"aecc01ef019089c0553d0114e17cae64e8ccc687c2b0b93b24e4e534ebfa9833"} Oct 10 16:53:11 crc kubenswrapper[4788]: I1010 16:53:11.790067 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" podStartSLOduration=2.626648788 podStartE2EDuration="2.790034873s" podCreationTimestamp="2025-10-10 16:53:09 +0000 UTC" firstStartedPulling="2025-10-10 16:53:10.866521715 +0000 UTC m=+7693.316237303" lastFinishedPulling="2025-10-10 16:53:11.02990783 +0000 UTC m=+7693.479623388" observedRunningTime="2025-10-10 16:53:11.775855353 +0000 UTC m=+7694.225570911" watchObservedRunningTime="2025-10-10 16:53:11.790034873 +0000 UTC m=+7694.239750431" Oct 10 16:53:25 crc kubenswrapper[4788]: I1010 16:53:25.234695 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:53:25 crc kubenswrapper[4788]: E1010 16:53:25.235609 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:53:30 crc kubenswrapper[4788]: I1010 16:53:30.973487 4788 generic.go:334] "Generic (PLEG): container finished" podID="6eff8747-d0bb-49d9-b45d-95ed5adf39a0" containerID="562d60a8b2233da9c94b2d639317a068489aa8b87a13854ad59eb0d802538d3d" exitCode=0 Oct 10 16:53:30 crc kubenswrapper[4788]: I1010 16:53:30.973740 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" event={"ID":"6eff8747-d0bb-49d9-b45d-95ed5adf39a0","Type":"ContainerDied","Data":"562d60a8b2233da9c94b2d639317a068489aa8b87a13854ad59eb0d802538d3d"} Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.484864 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575271 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-metadata-combined-ca-bundle\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575313 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-telemetry-combined-ca-bundle\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575397 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-sriov-combined-ca-bundle\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575449 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ceph\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575472 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-bootstrap-combined-ca-bundle\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575511 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-nova-combined-ca-bundle\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575545 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-libvirt-combined-ca-bundle\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575580 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ssh-key\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575645 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-dhcp-combined-ca-bundle\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575694 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ovn-combined-ca-bundle\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575720 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-inventory\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.575753 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65ngn\" (UniqueName: \"kubernetes.io/projected/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-kube-api-access-65ngn\") pod \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\" (UID: \"6eff8747-d0bb-49d9-b45d-95ed5adf39a0\") " Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.583243 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ceph" (OuterVolumeSpecName: "ceph") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.583293 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.583659 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.583687 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.583719 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-kube-api-access-65ngn" (OuterVolumeSpecName: "kube-api-access-65ngn") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "kube-api-access-65ngn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.592609 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.598957 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.598984 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.599039 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.599171 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.612275 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-inventory" (OuterVolumeSpecName: "inventory") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.616181 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6eff8747-d0bb-49d9-b45d-95ed5adf39a0" (UID: "6eff8747-d0bb-49d9-b45d-95ed5adf39a0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.705665 4788 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.705804 4788 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.705871 4788 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.705966 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.706045 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.706856 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.706932 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.707002 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65ngn\" (UniqueName: \"kubernetes.io/projected/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-kube-api-access-65ngn\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.707069 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.707168 4788 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.707252 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.707316 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6eff8747-d0bb-49d9-b45d-95ed5adf39a0-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.997466 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" event={"ID":"6eff8747-d0bb-49d9-b45d-95ed5adf39a0","Type":"ContainerDied","Data":"aecc01ef019089c0553d0114e17cae64e8ccc687c2b0b93b24e4e534ebfa9833"} Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.997784 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aecc01ef019089c0553d0114e17cae64e8ccc687c2b0b93b24e4e534ebfa9833" Oct 10 16:53:32 crc kubenswrapper[4788]: I1010 16:53:32.997727 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-97dq9" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.085313 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-gc2fn"] Oct 10 16:53:33 crc kubenswrapper[4788]: E1010 16:53:33.085744 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eff8747-d0bb-49d9-b45d-95ed5adf39a0" containerName="install-certs-openstack-openstack-cell1" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.085760 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eff8747-d0bb-49d9-b45d-95ed5adf39a0" containerName="install-certs-openstack-openstack-cell1" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.085981 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eff8747-d0bb-49d9-b45d-95ed5adf39a0" containerName="install-certs-openstack-openstack-cell1" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.086799 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.090700 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.092902 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.092995 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.096610 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.108112 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-gc2fn"] Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.217992 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.218122 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4tsl\" (UniqueName: \"kubernetes.io/projected/66598dac-a1d9-49ad-a8a6-f1d663d9907d-kube-api-access-z4tsl\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.218227 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-inventory\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.218333 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ceph\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.320409 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4tsl\" (UniqueName: \"kubernetes.io/projected/66598dac-a1d9-49ad-a8a6-f1d663d9907d-kube-api-access-z4tsl\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.320522 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-inventory\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.320750 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ceph\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.320871 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.325727 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ceph\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.326449 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.327485 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-inventory\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.349471 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4tsl\" (UniqueName: \"kubernetes.io/projected/66598dac-a1d9-49ad-a8a6-f1d663d9907d-kube-api-access-z4tsl\") pod \"ceph-client-openstack-openstack-cell1-gc2fn\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.410964 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:33 crc kubenswrapper[4788]: I1010 16:53:33.974814 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-gc2fn"] Oct 10 16:53:33 crc kubenswrapper[4788]: W1010 16:53:33.981222 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66598dac_a1d9_49ad_a8a6_f1d663d9907d.slice/crio-c8c2fbf4fef61078c45f302083b7e7d76ede080c813acd076d5574b69e3baade WatchSource:0}: Error finding container c8c2fbf4fef61078c45f302083b7e7d76ede080c813acd076d5574b69e3baade: Status 404 returned error can't find the container with id c8c2fbf4fef61078c45f302083b7e7d76ede080c813acd076d5574b69e3baade Oct 10 16:53:34 crc kubenswrapper[4788]: I1010 16:53:34.012743 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" event={"ID":"66598dac-a1d9-49ad-a8a6-f1d663d9907d","Type":"ContainerStarted","Data":"c8c2fbf4fef61078c45f302083b7e7d76ede080c813acd076d5574b69e3baade"} Oct 10 16:53:35 crc kubenswrapper[4788]: I1010 16:53:35.025395 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" event={"ID":"66598dac-a1d9-49ad-a8a6-f1d663d9907d","Type":"ContainerStarted","Data":"2a4b56dc2d99bbc8211a8b62c4845d49d007a1336961725bf2fb4af537f25558"} Oct 10 16:53:35 crc kubenswrapper[4788]: I1010 16:53:35.054607 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" podStartSLOduration=1.85234505 podStartE2EDuration="2.05458032s" podCreationTimestamp="2025-10-10 16:53:33 +0000 UTC" firstStartedPulling="2025-10-10 16:53:33.984357547 +0000 UTC m=+7716.434073095" lastFinishedPulling="2025-10-10 16:53:34.186592817 +0000 UTC m=+7716.636308365" observedRunningTime="2025-10-10 16:53:35.048958626 +0000 UTC m=+7717.498674224" watchObservedRunningTime="2025-10-10 16:53:35.05458032 +0000 UTC m=+7717.504295878" Oct 10 16:53:36 crc kubenswrapper[4788]: I1010 16:53:36.242200 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:53:36 crc kubenswrapper[4788]: E1010 16:53:36.247568 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:53:40 crc kubenswrapper[4788]: I1010 16:53:40.084221 4788 generic.go:334] "Generic (PLEG): container finished" podID="66598dac-a1d9-49ad-a8a6-f1d663d9907d" containerID="2a4b56dc2d99bbc8211a8b62c4845d49d007a1336961725bf2fb4af537f25558" exitCode=0 Oct 10 16:53:40 crc kubenswrapper[4788]: I1010 16:53:40.084314 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" event={"ID":"66598dac-a1d9-49ad-a8a6-f1d663d9907d","Type":"ContainerDied","Data":"2a4b56dc2d99bbc8211a8b62c4845d49d007a1336961725bf2fb4af537f25558"} Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.626800 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.826257 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4tsl\" (UniqueName: \"kubernetes.io/projected/66598dac-a1d9-49ad-a8a6-f1d663d9907d-kube-api-access-z4tsl\") pod \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.826465 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-inventory\") pod \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.826556 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ssh-key\") pod \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.826594 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ceph\") pod \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\" (UID: \"66598dac-a1d9-49ad-a8a6-f1d663d9907d\") " Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.833172 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ceph" (OuterVolumeSpecName: "ceph") pod "66598dac-a1d9-49ad-a8a6-f1d663d9907d" (UID: "66598dac-a1d9-49ad-a8a6-f1d663d9907d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.834954 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66598dac-a1d9-49ad-a8a6-f1d663d9907d-kube-api-access-z4tsl" (OuterVolumeSpecName: "kube-api-access-z4tsl") pod "66598dac-a1d9-49ad-a8a6-f1d663d9907d" (UID: "66598dac-a1d9-49ad-a8a6-f1d663d9907d"). InnerVolumeSpecName "kube-api-access-z4tsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.861228 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-inventory" (OuterVolumeSpecName: "inventory") pod "66598dac-a1d9-49ad-a8a6-f1d663d9907d" (UID: "66598dac-a1d9-49ad-a8a6-f1d663d9907d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.868017 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "66598dac-a1d9-49ad-a8a6-f1d663d9907d" (UID: "66598dac-a1d9-49ad-a8a6-f1d663d9907d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.931824 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.931885 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.931921 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/66598dac-a1d9-49ad-a8a6-f1d663d9907d-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:41 crc kubenswrapper[4788]: I1010 16:53:41.931942 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4tsl\" (UniqueName: \"kubernetes.io/projected/66598dac-a1d9-49ad-a8a6-f1d663d9907d-kube-api-access-z4tsl\") on node \"crc\" DevicePath \"\"" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.133133 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" event={"ID":"66598dac-a1d9-49ad-a8a6-f1d663d9907d","Type":"ContainerDied","Data":"c8c2fbf4fef61078c45f302083b7e7d76ede080c813acd076d5574b69e3baade"} Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.133220 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8c2fbf4fef61078c45f302083b7e7d76ede080c813acd076d5574b69e3baade" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.133310 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-gc2fn" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.297236 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-qbrs7"] Oct 10 16:53:42 crc kubenswrapper[4788]: E1010 16:53:42.298042 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66598dac-a1d9-49ad-a8a6-f1d663d9907d" containerName="ceph-client-openstack-openstack-cell1" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.298054 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="66598dac-a1d9-49ad-a8a6-f1d663d9907d" containerName="ceph-client-openstack-openstack-cell1" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.298334 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="66598dac-a1d9-49ad-a8a6-f1d663d9907d" containerName="ceph-client-openstack-openstack-cell1" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.299356 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.308547 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-qbrs7"] Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.313638 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.313841 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.314046 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.314250 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.317478 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.450622 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc21acf4-40c2-4cab-998e-5561eb853813-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.450762 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ceph\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.450807 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ssh-key\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.450831 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.450885 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvjnq\" (UniqueName: \"kubernetes.io/projected/bc21acf4-40c2-4cab-998e-5561eb853813-kube-api-access-hvjnq\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.450907 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-inventory\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.552859 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ceph\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.552922 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ssh-key\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.552948 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.553015 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvjnq\" (UniqueName: \"kubernetes.io/projected/bc21acf4-40c2-4cab-998e-5561eb853813-kube-api-access-hvjnq\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.553043 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-inventory\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.553089 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc21acf4-40c2-4cab-998e-5561eb853813-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.553977 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc21acf4-40c2-4cab-998e-5561eb853813-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.558713 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ssh-key\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.559808 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ceph\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.559846 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.560233 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-inventory\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.570922 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvjnq\" (UniqueName: \"kubernetes.io/projected/bc21acf4-40c2-4cab-998e-5561eb853813-kube-api-access-hvjnq\") pod \"ovn-openstack-openstack-cell1-qbrs7\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:42 crc kubenswrapper[4788]: I1010 16:53:42.643410 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:53:43 crc kubenswrapper[4788]: I1010 16:53:43.194513 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-qbrs7"] Oct 10 16:53:44 crc kubenswrapper[4788]: I1010 16:53:44.154179 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" event={"ID":"bc21acf4-40c2-4cab-998e-5561eb853813","Type":"ContainerStarted","Data":"68764d02bb9d3824ad0e2fc5b6f0a6bf3b7cb6d4b1fc2cb8d7e0979dc2e64e77"} Oct 10 16:53:44 crc kubenswrapper[4788]: I1010 16:53:44.155001 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" event={"ID":"bc21acf4-40c2-4cab-998e-5561eb853813","Type":"ContainerStarted","Data":"58f5a4edd8f00307ad63cb772765087464d0df81afcdac58af92902c338a302b"} Oct 10 16:53:44 crc kubenswrapper[4788]: I1010 16:53:44.177711 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" podStartSLOduration=2.003006692 podStartE2EDuration="2.177683397s" podCreationTimestamp="2025-10-10 16:53:42 +0000 UTC" firstStartedPulling="2025-10-10 16:53:43.198757049 +0000 UTC m=+7725.648472607" lastFinishedPulling="2025-10-10 16:53:43.373433764 +0000 UTC m=+7725.823149312" observedRunningTime="2025-10-10 16:53:44.171411906 +0000 UTC m=+7726.621127474" watchObservedRunningTime="2025-10-10 16:53:44.177683397 +0000 UTC m=+7726.627398955" Oct 10 16:53:47 crc kubenswrapper[4788]: I1010 16:53:47.234279 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:53:47 crc kubenswrapper[4788]: E1010 16:53:47.235193 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:53:59 crc kubenswrapper[4788]: I1010 16:53:59.233861 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:53:59 crc kubenswrapper[4788]: E1010 16:53:59.234885 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:54:13 crc kubenswrapper[4788]: I1010 16:54:13.234121 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:54:13 crc kubenswrapper[4788]: E1010 16:54:13.235230 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:54:24 crc kubenswrapper[4788]: I1010 16:54:24.235062 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:54:24 crc kubenswrapper[4788]: E1010 16:54:24.235863 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:54:37 crc kubenswrapper[4788]: I1010 16:54:37.234021 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:54:37 crc kubenswrapper[4788]: E1010 16:54:37.235272 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:54:48 crc kubenswrapper[4788]: I1010 16:54:48.244387 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:54:48 crc kubenswrapper[4788]: E1010 16:54:48.245284 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:54:51 crc kubenswrapper[4788]: I1010 16:54:51.927993 4788 generic.go:334] "Generic (PLEG): container finished" podID="bc21acf4-40c2-4cab-998e-5561eb853813" containerID="68764d02bb9d3824ad0e2fc5b6f0a6bf3b7cb6d4b1fc2cb8d7e0979dc2e64e77" exitCode=0 Oct 10 16:54:51 crc kubenswrapper[4788]: I1010 16:54:51.928226 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" event={"ID":"bc21acf4-40c2-4cab-998e-5561eb853813","Type":"ContainerDied","Data":"68764d02bb9d3824ad0e2fc5b6f0a6bf3b7cb6d4b1fc2cb8d7e0979dc2e64e77"} Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.476961 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.589253 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvjnq\" (UniqueName: \"kubernetes.io/projected/bc21acf4-40c2-4cab-998e-5561eb853813-kube-api-access-hvjnq\") pod \"bc21acf4-40c2-4cab-998e-5561eb853813\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.589473 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc21acf4-40c2-4cab-998e-5561eb853813-ovncontroller-config-0\") pod \"bc21acf4-40c2-4cab-998e-5561eb853813\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.589522 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ssh-key\") pod \"bc21acf4-40c2-4cab-998e-5561eb853813\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.589591 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-inventory\") pod \"bc21acf4-40c2-4cab-998e-5561eb853813\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.589699 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ovn-combined-ca-bundle\") pod \"bc21acf4-40c2-4cab-998e-5561eb853813\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.589773 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ceph\") pod \"bc21acf4-40c2-4cab-998e-5561eb853813\" (UID: \"bc21acf4-40c2-4cab-998e-5561eb853813\") " Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.595388 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bc21acf4-40c2-4cab-998e-5561eb853813" (UID: "bc21acf4-40c2-4cab-998e-5561eb853813"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.595456 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ceph" (OuterVolumeSpecName: "ceph") pod "bc21acf4-40c2-4cab-998e-5561eb853813" (UID: "bc21acf4-40c2-4cab-998e-5561eb853813"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.596000 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc21acf4-40c2-4cab-998e-5561eb853813-kube-api-access-hvjnq" (OuterVolumeSpecName: "kube-api-access-hvjnq") pod "bc21acf4-40c2-4cab-998e-5561eb853813" (UID: "bc21acf4-40c2-4cab-998e-5561eb853813"). InnerVolumeSpecName "kube-api-access-hvjnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.619765 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-inventory" (OuterVolumeSpecName: "inventory") pod "bc21acf4-40c2-4cab-998e-5561eb853813" (UID: "bc21acf4-40c2-4cab-998e-5561eb853813"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.638154 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc21acf4-40c2-4cab-998e-5561eb853813" (UID: "bc21acf4-40c2-4cab-998e-5561eb853813"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.645657 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc21acf4-40c2-4cab-998e-5561eb853813-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "bc21acf4-40c2-4cab-998e-5561eb853813" (UID: "bc21acf4-40c2-4cab-998e-5561eb853813"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.692775 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvjnq\" (UniqueName: \"kubernetes.io/projected/bc21acf4-40c2-4cab-998e-5561eb853813-kube-api-access-hvjnq\") on node \"crc\" DevicePath \"\"" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.693120 4788 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bc21acf4-40c2-4cab-998e-5561eb853813-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.693364 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.693514 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.693674 4788 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.693849 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bc21acf4-40c2-4cab-998e-5561eb853813-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.954104 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" event={"ID":"bc21acf4-40c2-4cab-998e-5561eb853813","Type":"ContainerDied","Data":"58f5a4edd8f00307ad63cb772765087464d0df81afcdac58af92902c338a302b"} Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.954651 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58f5a4edd8f00307ad63cb772765087464d0df81afcdac58af92902c338a302b" Oct 10 16:54:53 crc kubenswrapper[4788]: I1010 16:54:53.954185 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-qbrs7" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.058724 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-mgbkw"] Oct 10 16:54:54 crc kubenswrapper[4788]: E1010 16:54:54.059272 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc21acf4-40c2-4cab-998e-5561eb853813" containerName="ovn-openstack-openstack-cell1" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.059295 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc21acf4-40c2-4cab-998e-5561eb853813" containerName="ovn-openstack-openstack-cell1" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.059589 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc21acf4-40c2-4cab-998e-5561eb853813" containerName="ovn-openstack-openstack-cell1" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.060636 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.064169 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.064271 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.064282 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.065329 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.066555 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.072156 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.081133 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-mgbkw"] Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.210422 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.210659 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.210711 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.210737 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.210770 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.210802 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.210845 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svtp2\" (UniqueName: \"kubernetes.io/projected/83f651a1-892d-4422-964b-93c92ddfa185-kube-api-access-svtp2\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.314103 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.314264 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svtp2\" (UniqueName: \"kubernetes.io/projected/83f651a1-892d-4422-964b-93c92ddfa185-kube-api-access-svtp2\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.314394 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.314872 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.314978 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.315025 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.315096 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.318474 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.318997 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.321160 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.321393 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.322056 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.326384 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.333312 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svtp2\" (UniqueName: \"kubernetes.io/projected/83f651a1-892d-4422-964b-93c92ddfa185-kube-api-access-svtp2\") pod \"neutron-metadata-openstack-openstack-cell1-mgbkw\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.379353 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.897844 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-mgbkw"] Oct 10 16:54:54 crc kubenswrapper[4788]: I1010 16:54:54.967746 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" event={"ID":"83f651a1-892d-4422-964b-93c92ddfa185","Type":"ContainerStarted","Data":"58c8c121cdb195775c21cd387acc2ba69376d9e0d207bd308fb1ff91d26716b3"} Oct 10 16:54:55 crc kubenswrapper[4788]: I1010 16:54:55.979506 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" event={"ID":"83f651a1-892d-4422-964b-93c92ddfa185","Type":"ContainerStarted","Data":"e850f8379ddc23f17bf299d1f69de1bd14d10bc4ff21ed691cddda4b9b9b3d7c"} Oct 10 16:54:56 crc kubenswrapper[4788]: I1010 16:54:56.000607 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" podStartSLOduration=1.796790808 podStartE2EDuration="2.000587451s" podCreationTimestamp="2025-10-10 16:54:54 +0000 UTC" firstStartedPulling="2025-10-10 16:54:54.89731684 +0000 UTC m=+7797.347032388" lastFinishedPulling="2025-10-10 16:54:55.101113483 +0000 UTC m=+7797.550829031" observedRunningTime="2025-10-10 16:54:55.99948122 +0000 UTC m=+7798.449196768" watchObservedRunningTime="2025-10-10 16:54:56.000587451 +0000 UTC m=+7798.450302999" Oct 10 16:54:59 crc kubenswrapper[4788]: I1010 16:54:59.234614 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:54:59 crc kubenswrapper[4788]: E1010 16:54:59.235185 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:55:05 crc kubenswrapper[4788]: I1010 16:55:05.851027 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vg7bh"] Oct 10 16:55:05 crc kubenswrapper[4788]: I1010 16:55:05.859827 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:05 crc kubenswrapper[4788]: I1010 16:55:05.872798 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vg7bh"] Oct 10 16:55:05 crc kubenswrapper[4788]: I1010 16:55:05.955151 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blrxn\" (UniqueName: \"kubernetes.io/projected/2ee0384e-444c-4852-bb31-f6721b905a79-kube-api-access-blrxn\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:05 crc kubenswrapper[4788]: I1010 16:55:05.955632 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-utilities\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:05 crc kubenswrapper[4788]: I1010 16:55:05.955697 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-catalog-content\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:06 crc kubenswrapper[4788]: I1010 16:55:06.057794 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-catalog-content\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:06 crc kubenswrapper[4788]: I1010 16:55:06.057991 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blrxn\" (UniqueName: \"kubernetes.io/projected/2ee0384e-444c-4852-bb31-f6721b905a79-kube-api-access-blrxn\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:06 crc kubenswrapper[4788]: I1010 16:55:06.058060 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-utilities\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:06 crc kubenswrapper[4788]: I1010 16:55:06.058293 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-catalog-content\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:06 crc kubenswrapper[4788]: I1010 16:55:06.058606 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-utilities\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:06 crc kubenswrapper[4788]: I1010 16:55:06.077509 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blrxn\" (UniqueName: \"kubernetes.io/projected/2ee0384e-444c-4852-bb31-f6721b905a79-kube-api-access-blrxn\") pod \"certified-operators-vg7bh\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:06 crc kubenswrapper[4788]: I1010 16:55:06.222476 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:06 crc kubenswrapper[4788]: I1010 16:55:06.734796 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vg7bh"] Oct 10 16:55:07 crc kubenswrapper[4788]: I1010 16:55:07.125598 4788 generic.go:334] "Generic (PLEG): container finished" podID="2ee0384e-444c-4852-bb31-f6721b905a79" containerID="00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4" exitCode=0 Oct 10 16:55:07 crc kubenswrapper[4788]: I1010 16:55:07.125658 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vg7bh" event={"ID":"2ee0384e-444c-4852-bb31-f6721b905a79","Type":"ContainerDied","Data":"00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4"} Oct 10 16:55:07 crc kubenswrapper[4788]: I1010 16:55:07.125690 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vg7bh" event={"ID":"2ee0384e-444c-4852-bb31-f6721b905a79","Type":"ContainerStarted","Data":"6da40927126bc21629f77fb2a382ef49ed998a0b2327fdd194bb5297bb718de5"} Oct 10 16:55:08 crc kubenswrapper[4788]: I1010 16:55:08.141701 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vg7bh" event={"ID":"2ee0384e-444c-4852-bb31-f6721b905a79","Type":"ContainerStarted","Data":"7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b"} Oct 10 16:55:09 crc kubenswrapper[4788]: I1010 16:55:09.152198 4788 generic.go:334] "Generic (PLEG): container finished" podID="2ee0384e-444c-4852-bb31-f6721b905a79" containerID="7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b" exitCode=0 Oct 10 16:55:09 crc kubenswrapper[4788]: I1010 16:55:09.152307 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vg7bh" event={"ID":"2ee0384e-444c-4852-bb31-f6721b905a79","Type":"ContainerDied","Data":"7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b"} Oct 10 16:55:10 crc kubenswrapper[4788]: I1010 16:55:10.173575 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vg7bh" event={"ID":"2ee0384e-444c-4852-bb31-f6721b905a79","Type":"ContainerStarted","Data":"ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f"} Oct 10 16:55:10 crc kubenswrapper[4788]: I1010 16:55:10.229408 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vg7bh" podStartSLOduration=2.795077368 podStartE2EDuration="5.229385541s" podCreationTimestamp="2025-10-10 16:55:05 +0000 UTC" firstStartedPulling="2025-10-10 16:55:07.129366487 +0000 UTC m=+7809.579082035" lastFinishedPulling="2025-10-10 16:55:09.56367466 +0000 UTC m=+7812.013390208" observedRunningTime="2025-10-10 16:55:10.227051368 +0000 UTC m=+7812.676766916" watchObservedRunningTime="2025-10-10 16:55:10.229385541 +0000 UTC m=+7812.679101089" Oct 10 16:55:13 crc kubenswrapper[4788]: I1010 16:55:13.235312 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:55:13 crc kubenswrapper[4788]: E1010 16:55:13.236031 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:55:16 crc kubenswrapper[4788]: I1010 16:55:16.222662 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:16 crc kubenswrapper[4788]: I1010 16:55:16.223123 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:16 crc kubenswrapper[4788]: I1010 16:55:16.274174 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:16 crc kubenswrapper[4788]: I1010 16:55:16.339671 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:18 crc kubenswrapper[4788]: I1010 16:55:18.637570 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vg7bh"] Oct 10 16:55:18 crc kubenswrapper[4788]: I1010 16:55:18.638215 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vg7bh" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" containerName="registry-server" containerID="cri-o://ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f" gracePeriod=2 Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.150102 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.277405 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-catalog-content\") pod \"2ee0384e-444c-4852-bb31-f6721b905a79\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.277738 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-utilities\") pod \"2ee0384e-444c-4852-bb31-f6721b905a79\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.277809 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blrxn\" (UniqueName: \"kubernetes.io/projected/2ee0384e-444c-4852-bb31-f6721b905a79-kube-api-access-blrxn\") pod \"2ee0384e-444c-4852-bb31-f6721b905a79\" (UID: \"2ee0384e-444c-4852-bb31-f6721b905a79\") " Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.278814 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-utilities" (OuterVolumeSpecName: "utilities") pod "2ee0384e-444c-4852-bb31-f6721b905a79" (UID: "2ee0384e-444c-4852-bb31-f6721b905a79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.283198 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee0384e-444c-4852-bb31-f6721b905a79-kube-api-access-blrxn" (OuterVolumeSpecName: "kube-api-access-blrxn") pod "2ee0384e-444c-4852-bb31-f6721b905a79" (UID: "2ee0384e-444c-4852-bb31-f6721b905a79"). InnerVolumeSpecName "kube-api-access-blrxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.291806 4788 generic.go:334] "Generic (PLEG): container finished" podID="2ee0384e-444c-4852-bb31-f6721b905a79" containerID="ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f" exitCode=0 Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.292032 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vg7bh" event={"ID":"2ee0384e-444c-4852-bb31-f6721b905a79","Type":"ContainerDied","Data":"ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f"} Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.292190 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vg7bh" event={"ID":"2ee0384e-444c-4852-bb31-f6721b905a79","Type":"ContainerDied","Data":"6da40927126bc21629f77fb2a382ef49ed998a0b2327fdd194bb5297bb718de5"} Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.292055 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vg7bh" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.292302 4788 scope.go:117] "RemoveContainer" containerID="ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.324814 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ee0384e-444c-4852-bb31-f6721b905a79" (UID: "2ee0384e-444c-4852-bb31-f6721b905a79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.353459 4788 scope.go:117] "RemoveContainer" containerID="7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.377559 4788 scope.go:117] "RemoveContainer" containerID="00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.381568 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.381592 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee0384e-444c-4852-bb31-f6721b905a79-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.381606 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blrxn\" (UniqueName: \"kubernetes.io/projected/2ee0384e-444c-4852-bb31-f6721b905a79-kube-api-access-blrxn\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.440761 4788 scope.go:117] "RemoveContainer" containerID="ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f" Oct 10 16:55:19 crc kubenswrapper[4788]: E1010 16:55:19.441765 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f\": container with ID starting with ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f not found: ID does not exist" containerID="ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.441803 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f"} err="failed to get container status \"ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f\": rpc error: code = NotFound desc = could not find container \"ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f\": container with ID starting with ebd11500f83b375f2678739ed73841c96c68bc3e0fd2b672f1dda104b602439f not found: ID does not exist" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.441831 4788 scope.go:117] "RemoveContainer" containerID="7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b" Oct 10 16:55:19 crc kubenswrapper[4788]: E1010 16:55:19.442463 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b\": container with ID starting with 7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b not found: ID does not exist" containerID="7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.442576 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b"} err="failed to get container status \"7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b\": rpc error: code = NotFound desc = could not find container \"7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b\": container with ID starting with 7602c58d1ea93314a371759c784e2d0eb117db95bab5aa40b0ac83436f17439b not found: ID does not exist" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.442653 4788 scope.go:117] "RemoveContainer" containerID="00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4" Oct 10 16:55:19 crc kubenswrapper[4788]: E1010 16:55:19.442993 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4\": container with ID starting with 00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4 not found: ID does not exist" containerID="00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.443019 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4"} err="failed to get container status \"00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4\": rpc error: code = NotFound desc = could not find container \"00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4\": container with ID starting with 00e84d820608377a843a251c57660a00ccd47bcb2e9e67a757265632b63613a4 not found: ID does not exist" Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.635438 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vg7bh"] Oct 10 16:55:19 crc kubenswrapper[4788]: I1010 16:55:19.643960 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vg7bh"] Oct 10 16:55:20 crc kubenswrapper[4788]: I1010 16:55:20.258604 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" path="/var/lib/kubelet/pods/2ee0384e-444c-4852-bb31-f6721b905a79/volumes" Oct 10 16:55:25 crc kubenswrapper[4788]: I1010 16:55:25.234413 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:55:25 crc kubenswrapper[4788]: E1010 16:55:25.235687 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:55:38 crc kubenswrapper[4788]: I1010 16:55:38.244936 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:55:38 crc kubenswrapper[4788]: E1010 16:55:38.245705 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:55:48 crc kubenswrapper[4788]: I1010 16:55:48.631109 4788 generic.go:334] "Generic (PLEG): container finished" podID="83f651a1-892d-4422-964b-93c92ddfa185" containerID="e850f8379ddc23f17bf299d1f69de1bd14d10bc4ff21ed691cddda4b9b9b3d7c" exitCode=0 Oct 10 16:55:48 crc kubenswrapper[4788]: I1010 16:55:48.631180 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" event={"ID":"83f651a1-892d-4422-964b-93c92ddfa185","Type":"ContainerDied","Data":"e850f8379ddc23f17bf299d1f69de1bd14d10bc4ff21ed691cddda4b9b9b3d7c"} Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.148561 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.326090 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-metadata-combined-ca-bundle\") pod \"83f651a1-892d-4422-964b-93c92ddfa185\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.326134 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svtp2\" (UniqueName: \"kubernetes.io/projected/83f651a1-892d-4422-964b-93c92ddfa185-kube-api-access-svtp2\") pod \"83f651a1-892d-4422-964b-93c92ddfa185\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.326186 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-ovn-metadata-agent-neutron-config-0\") pod \"83f651a1-892d-4422-964b-93c92ddfa185\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.326389 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-inventory\") pod \"83f651a1-892d-4422-964b-93c92ddfa185\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.326425 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ssh-key\") pod \"83f651a1-892d-4422-964b-93c92ddfa185\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.326523 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-nova-metadata-neutron-config-0\") pod \"83f651a1-892d-4422-964b-93c92ddfa185\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.326564 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ceph\") pod \"83f651a1-892d-4422-964b-93c92ddfa185\" (UID: \"83f651a1-892d-4422-964b-93c92ddfa185\") " Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.332300 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "83f651a1-892d-4422-964b-93c92ddfa185" (UID: "83f651a1-892d-4422-964b-93c92ddfa185"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.332813 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f651a1-892d-4422-964b-93c92ddfa185-kube-api-access-svtp2" (OuterVolumeSpecName: "kube-api-access-svtp2") pod "83f651a1-892d-4422-964b-93c92ddfa185" (UID: "83f651a1-892d-4422-964b-93c92ddfa185"). InnerVolumeSpecName "kube-api-access-svtp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.333301 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ceph" (OuterVolumeSpecName: "ceph") pod "83f651a1-892d-4422-964b-93c92ddfa185" (UID: "83f651a1-892d-4422-964b-93c92ddfa185"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.361393 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "83f651a1-892d-4422-964b-93c92ddfa185" (UID: "83f651a1-892d-4422-964b-93c92ddfa185"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.363382 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "83f651a1-892d-4422-964b-93c92ddfa185" (UID: "83f651a1-892d-4422-964b-93c92ddfa185"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.368642 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-inventory" (OuterVolumeSpecName: "inventory") pod "83f651a1-892d-4422-964b-93c92ddfa185" (UID: "83f651a1-892d-4422-964b-93c92ddfa185"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.372092 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "83f651a1-892d-4422-964b-93c92ddfa185" (UID: "83f651a1-892d-4422-964b-93c92ddfa185"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.429945 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.430254 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svtp2\" (UniqueName: \"kubernetes.io/projected/83f651a1-892d-4422-964b-93c92ddfa185-kube-api-access-svtp2\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.430267 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.430280 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.430287 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.430296 4788 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.430327 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/83f651a1-892d-4422-964b-93c92ddfa185-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.660675 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" event={"ID":"83f651a1-892d-4422-964b-93c92ddfa185","Type":"ContainerDied","Data":"58c8c121cdb195775c21cd387acc2ba69376d9e0d207bd308fb1ff91d26716b3"} Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.660745 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58c8c121cdb195775c21cd387acc2ba69376d9e0d207bd308fb1ff91d26716b3" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.660769 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-mgbkw" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.756435 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-47fvb"] Oct 10 16:55:50 crc kubenswrapper[4788]: E1010 16:55:50.756967 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" containerName="extract-utilities" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.757062 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" containerName="extract-utilities" Oct 10 16:55:50 crc kubenswrapper[4788]: E1010 16:55:50.757078 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f651a1-892d-4422-964b-93c92ddfa185" containerName="neutron-metadata-openstack-openstack-cell1" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.757085 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f651a1-892d-4422-964b-93c92ddfa185" containerName="neutron-metadata-openstack-openstack-cell1" Oct 10 16:55:50 crc kubenswrapper[4788]: E1010 16:55:50.757127 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" containerName="extract-content" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.757135 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" containerName="extract-content" Oct 10 16:55:50 crc kubenswrapper[4788]: E1010 16:55:50.757155 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" containerName="registry-server" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.757164 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" containerName="registry-server" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.757418 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee0384e-444c-4852-bb31-f6721b905a79" containerName="registry-server" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.757450 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f651a1-892d-4422-964b-93c92ddfa185" containerName="neutron-metadata-openstack-openstack-cell1" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.758649 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.761079 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.763878 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.763963 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.764081 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.764125 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.793128 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-47fvb"] Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.944278 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-inventory\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.944335 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.944379 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtsvt\" (UniqueName: \"kubernetes.io/projected/15644ee5-8d10-47e1-a124-a7df41949cda-kube-api-access-gtsvt\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.944712 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ceph\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.945071 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:50 crc kubenswrapper[4788]: I1010 16:55:50.945121 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ssh-key\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.047779 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.047872 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ssh-key\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.047969 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-inventory\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.048012 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.048069 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtsvt\" (UniqueName: \"kubernetes.io/projected/15644ee5-8d10-47e1-a124-a7df41949cda-kube-api-access-gtsvt\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.048176 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ceph\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.052444 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.052968 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-inventory\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.053301 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ssh-key\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.053947 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ceph\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.058463 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.072830 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtsvt\" (UniqueName: \"kubernetes.io/projected/15644ee5-8d10-47e1-a124-a7df41949cda-kube-api-access-gtsvt\") pod \"libvirt-openstack-openstack-cell1-47fvb\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:51 crc kubenswrapper[4788]: I1010 16:55:51.378568 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 16:55:52 crc kubenswrapper[4788]: I1010 16:55:52.075210 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-47fvb"] Oct 10 16:55:52 crc kubenswrapper[4788]: I1010 16:55:52.098445 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 16:55:52 crc kubenswrapper[4788]: I1010 16:55:52.695193 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" event={"ID":"15644ee5-8d10-47e1-a124-a7df41949cda","Type":"ContainerStarted","Data":"ecf420db8dec16063f8f5ee798a9f0b9fece40717ff6309b56edb914b1a24726"} Oct 10 16:55:52 crc kubenswrapper[4788]: I1010 16:55:52.696918 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" event={"ID":"15644ee5-8d10-47e1-a124-a7df41949cda","Type":"ContainerStarted","Data":"a2dfbf8666177f0f91e665a0b461fb22e78e4148ae422679d4a17b9363e5c99e"} Oct 10 16:55:52 crc kubenswrapper[4788]: I1010 16:55:52.725480 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" podStartSLOduration=2.525714276 podStartE2EDuration="2.725458118s" podCreationTimestamp="2025-10-10 16:55:50 +0000 UTC" firstStartedPulling="2025-10-10 16:55:52.098084358 +0000 UTC m=+7854.547799916" lastFinishedPulling="2025-10-10 16:55:52.29782821 +0000 UTC m=+7854.747543758" observedRunningTime="2025-10-10 16:55:52.718107525 +0000 UTC m=+7855.167823073" watchObservedRunningTime="2025-10-10 16:55:52.725458118 +0000 UTC m=+7855.175173666" Oct 10 16:55:53 crc kubenswrapper[4788]: I1010 16:55:53.235066 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:55:53 crc kubenswrapper[4788]: E1010 16:55:53.235385 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:56:04 crc kubenswrapper[4788]: I1010 16:56:04.234562 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:56:04 crc kubenswrapper[4788]: E1010 16:56:04.235276 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:56:18 crc kubenswrapper[4788]: I1010 16:56:18.253963 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:56:18 crc kubenswrapper[4788]: E1010 16:56:18.255856 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:56:30 crc kubenswrapper[4788]: I1010 16:56:30.234545 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:56:30 crc kubenswrapper[4788]: E1010 16:56:30.235425 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:56:43 crc kubenswrapper[4788]: I1010 16:56:43.234625 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:56:43 crc kubenswrapper[4788]: E1010 16:56:43.235998 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:56:55 crc kubenswrapper[4788]: I1010 16:56:55.234673 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:56:55 crc kubenswrapper[4788]: E1010 16:56:55.235609 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:57:07 crc kubenswrapper[4788]: I1010 16:57:07.234940 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:57:07 crc kubenswrapper[4788]: E1010 16:57:07.236231 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:57:20 crc kubenswrapper[4788]: I1010 16:57:20.236685 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:57:20 crc kubenswrapper[4788]: E1010 16:57:20.238282 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:57:32 crc kubenswrapper[4788]: I1010 16:57:32.237213 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:57:32 crc kubenswrapper[4788]: E1010 16:57:32.238430 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:57:44 crc kubenswrapper[4788]: I1010 16:57:44.234565 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:57:44 crc kubenswrapper[4788]: E1010 16:57:44.237747 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:57:56 crc kubenswrapper[4788]: I1010 16:57:56.236444 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:57:56 crc kubenswrapper[4788]: E1010 16:57:56.237581 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 16:58:11 crc kubenswrapper[4788]: I1010 16:58:11.234378 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 16:58:12 crc kubenswrapper[4788]: I1010 16:58:12.412604 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"2827f4232c96f62563197d686ef21e1ef96a8c95d1b753f0c3a009ab743fbaeb"} Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.010013 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bn8gl"] Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.015381 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.025832 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bn8gl"] Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.154715 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-catalog-content\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.155486 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-utilities\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.155979 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k9hb\" (UniqueName: \"kubernetes.io/projected/0ddbd5a5-ee90-4557-8513-6583f73e89b0-kube-api-access-5k9hb\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.258351 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k9hb\" (UniqueName: \"kubernetes.io/projected/0ddbd5a5-ee90-4557-8513-6583f73e89b0-kube-api-access-5k9hb\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.258541 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-catalog-content\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.258663 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-utilities\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.259559 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-catalog-content\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.259721 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-utilities\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.292438 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k9hb\" (UniqueName: \"kubernetes.io/projected/0ddbd5a5-ee90-4557-8513-6583f73e89b0-kube-api-access-5k9hb\") pod \"community-operators-bn8gl\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.376788 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:27 crc kubenswrapper[4788]: I1010 16:59:27.942007 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bn8gl"] Oct 10 16:59:28 crc kubenswrapper[4788]: I1010 16:59:28.210859 4788 generic.go:334] "Generic (PLEG): container finished" podID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerID="d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d" exitCode=0 Oct 10 16:59:28 crc kubenswrapper[4788]: I1010 16:59:28.210915 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bn8gl" event={"ID":"0ddbd5a5-ee90-4557-8513-6583f73e89b0","Type":"ContainerDied","Data":"d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d"} Oct 10 16:59:28 crc kubenswrapper[4788]: I1010 16:59:28.211279 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bn8gl" event={"ID":"0ddbd5a5-ee90-4557-8513-6583f73e89b0","Type":"ContainerStarted","Data":"590b3ae5194d1dc057398f9d9cde132bdd875b6e5faa105ca58f87ff1fb766c9"} Oct 10 16:59:29 crc kubenswrapper[4788]: I1010 16:59:29.224679 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bn8gl" event={"ID":"0ddbd5a5-ee90-4557-8513-6583f73e89b0","Type":"ContainerStarted","Data":"3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897"} Oct 10 16:59:30 crc kubenswrapper[4788]: I1010 16:59:30.245710 4788 generic.go:334] "Generic (PLEG): container finished" podID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerID="3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897" exitCode=0 Oct 10 16:59:30 crc kubenswrapper[4788]: I1010 16:59:30.256259 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bn8gl" event={"ID":"0ddbd5a5-ee90-4557-8513-6583f73e89b0","Type":"ContainerDied","Data":"3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897"} Oct 10 16:59:31 crc kubenswrapper[4788]: I1010 16:59:31.264401 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bn8gl" event={"ID":"0ddbd5a5-ee90-4557-8513-6583f73e89b0","Type":"ContainerStarted","Data":"376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514"} Oct 10 16:59:31 crc kubenswrapper[4788]: I1010 16:59:31.290299 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bn8gl" podStartSLOduration=2.8024575499999997 podStartE2EDuration="5.290273361s" podCreationTimestamp="2025-10-10 16:59:26 +0000 UTC" firstStartedPulling="2025-10-10 16:59:28.212988104 +0000 UTC m=+8070.662703652" lastFinishedPulling="2025-10-10 16:59:30.700803905 +0000 UTC m=+8073.150519463" observedRunningTime="2025-10-10 16:59:31.281501249 +0000 UTC m=+8073.731216797" watchObservedRunningTime="2025-10-10 16:59:31.290273361 +0000 UTC m=+8073.739988909" Oct 10 16:59:37 crc kubenswrapper[4788]: I1010 16:59:37.377120 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:37 crc kubenswrapper[4788]: I1010 16:59:37.378531 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:37 crc kubenswrapper[4788]: I1010 16:59:37.456903 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:38 crc kubenswrapper[4788]: I1010 16:59:38.404804 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:38 crc kubenswrapper[4788]: I1010 16:59:38.464671 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bn8gl"] Oct 10 16:59:40 crc kubenswrapper[4788]: I1010 16:59:40.421569 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bn8gl" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerName="registry-server" containerID="cri-o://376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514" gracePeriod=2 Oct 10 16:59:40 crc kubenswrapper[4788]: I1010 16:59:40.947878 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.034027 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k9hb\" (UniqueName: \"kubernetes.io/projected/0ddbd5a5-ee90-4557-8513-6583f73e89b0-kube-api-access-5k9hb\") pod \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.034167 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-catalog-content\") pod \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.034566 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-utilities\") pod \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\" (UID: \"0ddbd5a5-ee90-4557-8513-6583f73e89b0\") " Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.039657 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-utilities" (OuterVolumeSpecName: "utilities") pod "0ddbd5a5-ee90-4557-8513-6583f73e89b0" (UID: "0ddbd5a5-ee90-4557-8513-6583f73e89b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.062463 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ddbd5a5-ee90-4557-8513-6583f73e89b0-kube-api-access-5k9hb" (OuterVolumeSpecName: "kube-api-access-5k9hb") pod "0ddbd5a5-ee90-4557-8513-6583f73e89b0" (UID: "0ddbd5a5-ee90-4557-8513-6583f73e89b0"). InnerVolumeSpecName "kube-api-access-5k9hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.090722 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ddbd5a5-ee90-4557-8513-6583f73e89b0" (UID: "0ddbd5a5-ee90-4557-8513-6583f73e89b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.138995 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k9hb\" (UniqueName: \"kubernetes.io/projected/0ddbd5a5-ee90-4557-8513-6583f73e89b0-kube-api-access-5k9hb\") on node \"crc\" DevicePath \"\"" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.139045 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.139059 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddbd5a5-ee90-4557-8513-6583f73e89b0-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.436034 4788 generic.go:334] "Generic (PLEG): container finished" podID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerID="376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514" exitCode=0 Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.436090 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bn8gl" event={"ID":"0ddbd5a5-ee90-4557-8513-6583f73e89b0","Type":"ContainerDied","Data":"376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514"} Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.436124 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bn8gl" event={"ID":"0ddbd5a5-ee90-4557-8513-6583f73e89b0","Type":"ContainerDied","Data":"590b3ae5194d1dc057398f9d9cde132bdd875b6e5faa105ca58f87ff1fb766c9"} Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.436132 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bn8gl" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.436178 4788 scope.go:117] "RemoveContainer" containerID="376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.466060 4788 scope.go:117] "RemoveContainer" containerID="3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.504688 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bn8gl"] Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.519657 4788 scope.go:117] "RemoveContainer" containerID="d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.539128 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bn8gl"] Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.567303 4788 scope.go:117] "RemoveContainer" containerID="376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514" Oct 10 16:59:41 crc kubenswrapper[4788]: E1010 16:59:41.570075 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514\": container with ID starting with 376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514 not found: ID does not exist" containerID="376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.570127 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514"} err="failed to get container status \"376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514\": rpc error: code = NotFound desc = could not find container \"376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514\": container with ID starting with 376637401fd4ddb365e7cbaf81ab86f1379c29e4ac8357ea3f71be157ad9f514 not found: ID does not exist" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.570172 4788 scope.go:117] "RemoveContainer" containerID="3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897" Oct 10 16:59:41 crc kubenswrapper[4788]: E1010 16:59:41.572903 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897\": container with ID starting with 3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897 not found: ID does not exist" containerID="3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.572960 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897"} err="failed to get container status \"3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897\": rpc error: code = NotFound desc = could not find container \"3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897\": container with ID starting with 3455f0acbe0f69dd5817cd468d213c349bf0236e545e02afd662a5eac7af1897 not found: ID does not exist" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.572988 4788 scope.go:117] "RemoveContainer" containerID="d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d" Oct 10 16:59:41 crc kubenswrapper[4788]: E1010 16:59:41.574265 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d\": container with ID starting with d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d not found: ID does not exist" containerID="d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d" Oct 10 16:59:41 crc kubenswrapper[4788]: I1010 16:59:41.574323 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d"} err="failed to get container status \"d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d\": rpc error: code = NotFound desc = could not find container \"d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d\": container with ID starting with d5d471d9b835dfa3ce023f9a84973c1e8b8d5ad4934c887279283fd8b4fe443d not found: ID does not exist" Oct 10 16:59:42 crc kubenswrapper[4788]: I1010 16:59:42.253827 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" path="/var/lib/kubelet/pods/0ddbd5a5-ee90-4557-8513-6583f73e89b0/volumes" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.154623 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt"] Oct 10 17:00:00 crc kubenswrapper[4788]: E1010 17:00:00.155756 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerName="registry-server" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.155769 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerName="registry-server" Oct 10 17:00:00 crc kubenswrapper[4788]: E1010 17:00:00.155786 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerName="extract-utilities" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.155812 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerName="extract-utilities" Oct 10 17:00:00 crc kubenswrapper[4788]: E1010 17:00:00.155848 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerName="extract-content" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.155853 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerName="extract-content" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.156135 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ddbd5a5-ee90-4557-8513-6583f73e89b0" containerName="registry-server" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.157041 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.161251 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.162765 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.188795 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt"] Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.270345 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e827612e-795d-44b0-8fe2-1e17af0b95a3-secret-volume\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.270489 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s9h5\" (UniqueName: \"kubernetes.io/projected/e827612e-795d-44b0-8fe2-1e17af0b95a3-kube-api-access-9s9h5\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.270562 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e827612e-795d-44b0-8fe2-1e17af0b95a3-config-volume\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.372424 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e827612e-795d-44b0-8fe2-1e17af0b95a3-config-volume\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.372559 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e827612e-795d-44b0-8fe2-1e17af0b95a3-secret-volume\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.372669 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s9h5\" (UniqueName: \"kubernetes.io/projected/e827612e-795d-44b0-8fe2-1e17af0b95a3-kube-api-access-9s9h5\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.373863 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e827612e-795d-44b0-8fe2-1e17af0b95a3-config-volume\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.380758 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e827612e-795d-44b0-8fe2-1e17af0b95a3-secret-volume\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.390829 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s9h5\" (UniqueName: \"kubernetes.io/projected/e827612e-795d-44b0-8fe2-1e17af0b95a3-kube-api-access-9s9h5\") pod \"collect-profiles-29335260-kqpkt\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.501530 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:00 crc kubenswrapper[4788]: I1010 17:00:00.993850 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt"] Oct 10 17:00:01 crc kubenswrapper[4788]: I1010 17:00:01.649808 4788 generic.go:334] "Generic (PLEG): container finished" podID="e827612e-795d-44b0-8fe2-1e17af0b95a3" containerID="ba48650a8446ad57a2e48aa27528706885877e99e7a40f8e791513745a301c9c" exitCode=0 Oct 10 17:00:01 crc kubenswrapper[4788]: I1010 17:00:01.649913 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" event={"ID":"e827612e-795d-44b0-8fe2-1e17af0b95a3","Type":"ContainerDied","Data":"ba48650a8446ad57a2e48aa27528706885877e99e7a40f8e791513745a301c9c"} Oct 10 17:00:01 crc kubenswrapper[4788]: I1010 17:00:01.650311 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" event={"ID":"e827612e-795d-44b0-8fe2-1e17af0b95a3","Type":"ContainerStarted","Data":"aabda2f83d626b6f7228c53f9d04b82e81aae200fff99d0b5d01fea4a6ae8c87"} Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.117216 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.240947 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e827612e-795d-44b0-8fe2-1e17af0b95a3-secret-volume\") pod \"e827612e-795d-44b0-8fe2-1e17af0b95a3\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.240999 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e827612e-795d-44b0-8fe2-1e17af0b95a3-config-volume\") pod \"e827612e-795d-44b0-8fe2-1e17af0b95a3\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.241067 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s9h5\" (UniqueName: \"kubernetes.io/projected/e827612e-795d-44b0-8fe2-1e17af0b95a3-kube-api-access-9s9h5\") pod \"e827612e-795d-44b0-8fe2-1e17af0b95a3\" (UID: \"e827612e-795d-44b0-8fe2-1e17af0b95a3\") " Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.242063 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e827612e-795d-44b0-8fe2-1e17af0b95a3-config-volume" (OuterVolumeSpecName: "config-volume") pod "e827612e-795d-44b0-8fe2-1e17af0b95a3" (UID: "e827612e-795d-44b0-8fe2-1e17af0b95a3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.248353 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e827612e-795d-44b0-8fe2-1e17af0b95a3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e827612e-795d-44b0-8fe2-1e17af0b95a3" (UID: "e827612e-795d-44b0-8fe2-1e17af0b95a3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.248632 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e827612e-795d-44b0-8fe2-1e17af0b95a3-kube-api-access-9s9h5" (OuterVolumeSpecName: "kube-api-access-9s9h5") pod "e827612e-795d-44b0-8fe2-1e17af0b95a3" (UID: "e827612e-795d-44b0-8fe2-1e17af0b95a3"). InnerVolumeSpecName "kube-api-access-9s9h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.344179 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e827612e-795d-44b0-8fe2-1e17af0b95a3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.344219 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e827612e-795d-44b0-8fe2-1e17af0b95a3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.344232 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s9h5\" (UniqueName: \"kubernetes.io/projected/e827612e-795d-44b0-8fe2-1e17af0b95a3-kube-api-access-9s9h5\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.684839 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" event={"ID":"e827612e-795d-44b0-8fe2-1e17af0b95a3","Type":"ContainerDied","Data":"aabda2f83d626b6f7228c53f9d04b82e81aae200fff99d0b5d01fea4a6ae8c87"} Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.684889 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aabda2f83d626b6f7228c53f9d04b82e81aae200fff99d0b5d01fea4a6ae8c87" Oct 10 17:00:03 crc kubenswrapper[4788]: I1010 17:00:03.684920 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335260-kqpkt" Oct 10 17:00:04 crc kubenswrapper[4788]: I1010 17:00:04.253033 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff"] Oct 10 17:00:04 crc kubenswrapper[4788]: I1010 17:00:04.255959 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335215-vdkff"] Oct 10 17:00:06 crc kubenswrapper[4788]: I1010 17:00:06.253045 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b54db3b-37c4-45b5-a0ab-b753c63244d2" path="/var/lib/kubelet/pods/5b54db3b-37c4-45b5-a0ab-b753c63244d2/volumes" Oct 10 17:00:20 crc kubenswrapper[4788]: I1010 17:00:20.363285 4788 scope.go:117] "RemoveContainer" containerID="d055b2a89ac62efd456f6596a77a1db560101832dead2e6f6300d201c22ab46d" Oct 10 17:00:29 crc kubenswrapper[4788]: I1010 17:00:29.407000 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:00:29 crc kubenswrapper[4788]: I1010 17:00:29.407729 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:00:35 crc kubenswrapper[4788]: I1010 17:00:35.023927 4788 generic.go:334] "Generic (PLEG): container finished" podID="15644ee5-8d10-47e1-a124-a7df41949cda" containerID="ecf420db8dec16063f8f5ee798a9f0b9fece40717ff6309b56edb914b1a24726" exitCode=0 Oct 10 17:00:35 crc kubenswrapper[4788]: I1010 17:00:35.024006 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" event={"ID":"15644ee5-8d10-47e1-a124-a7df41949cda","Type":"ContainerDied","Data":"ecf420db8dec16063f8f5ee798a9f0b9fece40717ff6309b56edb914b1a24726"} Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.553541 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.589378 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ssh-key\") pod \"15644ee5-8d10-47e1-a124-a7df41949cda\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.589463 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtsvt\" (UniqueName: \"kubernetes.io/projected/15644ee5-8d10-47e1-a124-a7df41949cda-kube-api-access-gtsvt\") pod \"15644ee5-8d10-47e1-a124-a7df41949cda\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.589549 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ceph\") pod \"15644ee5-8d10-47e1-a124-a7df41949cda\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.589811 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-secret-0\") pod \"15644ee5-8d10-47e1-a124-a7df41949cda\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.589841 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-inventory\") pod \"15644ee5-8d10-47e1-a124-a7df41949cda\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.589919 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-combined-ca-bundle\") pod \"15644ee5-8d10-47e1-a124-a7df41949cda\" (UID: \"15644ee5-8d10-47e1-a124-a7df41949cda\") " Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.596497 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ceph" (OuterVolumeSpecName: "ceph") pod "15644ee5-8d10-47e1-a124-a7df41949cda" (UID: "15644ee5-8d10-47e1-a124-a7df41949cda"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.598051 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.598866 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15644ee5-8d10-47e1-a124-a7df41949cda-kube-api-access-gtsvt" (OuterVolumeSpecName: "kube-api-access-gtsvt") pod "15644ee5-8d10-47e1-a124-a7df41949cda" (UID: "15644ee5-8d10-47e1-a124-a7df41949cda"). InnerVolumeSpecName "kube-api-access-gtsvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.602533 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "15644ee5-8d10-47e1-a124-a7df41949cda" (UID: "15644ee5-8d10-47e1-a124-a7df41949cda"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.624831 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "15644ee5-8d10-47e1-a124-a7df41949cda" (UID: "15644ee5-8d10-47e1-a124-a7df41949cda"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.631198 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "15644ee5-8d10-47e1-a124-a7df41949cda" (UID: "15644ee5-8d10-47e1-a124-a7df41949cda"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.632287 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-inventory" (OuterVolumeSpecName: "inventory") pod "15644ee5-8d10-47e1-a124-a7df41949cda" (UID: "15644ee5-8d10-47e1-a124-a7df41949cda"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.701809 4788 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.702033 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.702334 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtsvt\" (UniqueName: \"kubernetes.io/projected/15644ee5-8d10-47e1-a124-a7df41949cda-kube-api-access-gtsvt\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.702474 4788 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:36 crc kubenswrapper[4788]: I1010 17:00:36.702597 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15644ee5-8d10-47e1-a124-a7df41949cda-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.045663 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" event={"ID":"15644ee5-8d10-47e1-a124-a7df41949cda","Type":"ContainerDied","Data":"a2dfbf8666177f0f91e665a0b461fb22e78e4148ae422679d4a17b9363e5c99e"} Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.045732 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2dfbf8666177f0f91e665a0b461fb22e78e4148ae422679d4a17b9363e5c99e" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.045747 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-47fvb" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.163986 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-cjc4m"] Oct 10 17:00:37 crc kubenswrapper[4788]: E1010 17:00:37.165513 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e827612e-795d-44b0-8fe2-1e17af0b95a3" containerName="collect-profiles" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.165531 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e827612e-795d-44b0-8fe2-1e17af0b95a3" containerName="collect-profiles" Oct 10 17:00:37 crc kubenswrapper[4788]: E1010 17:00:37.165558 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15644ee5-8d10-47e1-a124-a7df41949cda" containerName="libvirt-openstack-openstack-cell1" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.165566 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="15644ee5-8d10-47e1-a124-a7df41949cda" containerName="libvirt-openstack-openstack-cell1" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.167032 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e827612e-795d-44b0-8fe2-1e17af0b95a3" containerName="collect-profiles" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.167081 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="15644ee5-8d10-47e1-a124-a7df41949cda" containerName="libvirt-openstack-openstack-cell1" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.170133 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.175357 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.176273 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.177981 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.178923 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.179843 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.179868 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.182461 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.206292 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-cjc4m"] Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221087 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ceph\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221201 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221259 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221289 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221318 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221341 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221357 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221377 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221392 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t7cz\" (UniqueName: \"kubernetes.io/projected/e1c3f042-7a1a-4a21-accd-636a15e88010-kube-api-access-2t7cz\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221411 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.221500 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-inventory\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.323833 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-inventory\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.323886 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ceph\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.323965 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.324065 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.324091 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.324119 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.324146 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.324193 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.324218 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.324234 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t7cz\" (UniqueName: \"kubernetes.io/projected/e1c3f042-7a1a-4a21-accd-636a15e88010-kube-api-access-2t7cz\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.324256 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.326505 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.326711 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.329906 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-inventory\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.330037 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ceph\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.330191 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.330827 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.331231 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.331867 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.336670 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.341013 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.346156 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t7cz\" (UniqueName: \"kubernetes.io/projected/e1c3f042-7a1a-4a21-accd-636a15e88010-kube-api-access-2t7cz\") pod \"nova-cell1-openstack-openstack-cell1-cjc4m\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:37 crc kubenswrapper[4788]: I1010 17:00:37.494122 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:00:38 crc kubenswrapper[4788]: I1010 17:00:38.076803 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-cjc4m"] Oct 10 17:00:39 crc kubenswrapper[4788]: I1010 17:00:39.065640 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" event={"ID":"e1c3f042-7a1a-4a21-accd-636a15e88010","Type":"ContainerStarted","Data":"13ea88933c752f55224fc19c564aefa885eb7fef401ccbf63d0a8b3b392e7239"} Oct 10 17:00:39 crc kubenswrapper[4788]: I1010 17:00:39.065986 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" event={"ID":"e1c3f042-7a1a-4a21-accd-636a15e88010","Type":"ContainerStarted","Data":"5015a2289475291125302c96331d84250dbe69c4f0415661604feee7afdbb2cf"} Oct 10 17:00:39 crc kubenswrapper[4788]: I1010 17:00:39.098261 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" podStartSLOduration=1.876209121 podStartE2EDuration="2.09824154s" podCreationTimestamp="2025-10-10 17:00:37 +0000 UTC" firstStartedPulling="2025-10-10 17:00:38.079157939 +0000 UTC m=+8140.528873487" lastFinishedPulling="2025-10-10 17:00:38.301190358 +0000 UTC m=+8140.750905906" observedRunningTime="2025-10-10 17:00:39.090896785 +0000 UTC m=+8141.540612323" watchObservedRunningTime="2025-10-10 17:00:39.09824154 +0000 UTC m=+8141.547957088" Oct 10 17:00:59 crc kubenswrapper[4788]: I1010 17:00:59.406447 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:00:59 crc kubenswrapper[4788]: I1010 17:00:59.406969 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.163365 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29335261-dvp6q"] Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.165348 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.182929 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29335261-dvp6q"] Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.269836 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-config-data\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.269929 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2klb\" (UniqueName: \"kubernetes.io/projected/d252c223-b923-4821-a4ff-8c183277565d-kube-api-access-p2klb\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.269983 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-combined-ca-bundle\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.270316 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-fernet-keys\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.372857 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-config-data\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.372978 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2klb\" (UniqueName: \"kubernetes.io/projected/d252c223-b923-4821-a4ff-8c183277565d-kube-api-access-p2klb\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.373010 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-combined-ca-bundle\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.373085 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-fernet-keys\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.381812 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-config-data\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.382251 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-combined-ca-bundle\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.383658 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-fernet-keys\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.392950 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2klb\" (UniqueName: \"kubernetes.io/projected/d252c223-b923-4821-a4ff-8c183277565d-kube-api-access-p2klb\") pod \"keystone-cron-29335261-dvp6q\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:00 crc kubenswrapper[4788]: I1010 17:01:00.501016 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:01 crc kubenswrapper[4788]: I1010 17:01:01.011480 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29335261-dvp6q"] Oct 10 17:01:01 crc kubenswrapper[4788]: W1010 17:01:01.017714 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd252c223_b923_4821_a4ff_8c183277565d.slice/crio-c9f5dc4a4610c278738f5dd22f18490aa4be03aa70693c71288487ccfb4252ac WatchSource:0}: Error finding container c9f5dc4a4610c278738f5dd22f18490aa4be03aa70693c71288487ccfb4252ac: Status 404 returned error can't find the container with id c9f5dc4a4610c278738f5dd22f18490aa4be03aa70693c71288487ccfb4252ac Oct 10 17:01:01 crc kubenswrapper[4788]: I1010 17:01:01.324383 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335261-dvp6q" event={"ID":"d252c223-b923-4821-a4ff-8c183277565d","Type":"ContainerStarted","Data":"a729b4f0599601263006722c731f267a4cbccb623d29de21df224afc8b451b54"} Oct 10 17:01:01 crc kubenswrapper[4788]: I1010 17:01:01.325294 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335261-dvp6q" event={"ID":"d252c223-b923-4821-a4ff-8c183277565d","Type":"ContainerStarted","Data":"c9f5dc4a4610c278738f5dd22f18490aa4be03aa70693c71288487ccfb4252ac"} Oct 10 17:01:01 crc kubenswrapper[4788]: I1010 17:01:01.353960 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29335261-dvp6q" podStartSLOduration=1.353934722 podStartE2EDuration="1.353934722s" podCreationTimestamp="2025-10-10 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 17:01:01.342357544 +0000 UTC m=+8163.792073132" watchObservedRunningTime="2025-10-10 17:01:01.353934722 +0000 UTC m=+8163.803650270" Oct 10 17:01:04 crc kubenswrapper[4788]: I1010 17:01:04.356510 4788 generic.go:334] "Generic (PLEG): container finished" podID="d252c223-b923-4821-a4ff-8c183277565d" containerID="a729b4f0599601263006722c731f267a4cbccb623d29de21df224afc8b451b54" exitCode=0 Oct 10 17:01:04 crc kubenswrapper[4788]: I1010 17:01:04.356681 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335261-dvp6q" event={"ID":"d252c223-b923-4821-a4ff-8c183277565d","Type":"ContainerDied","Data":"a729b4f0599601263006722c731f267a4cbccb623d29de21df224afc8b451b54"} Oct 10 17:01:05 crc kubenswrapper[4788]: I1010 17:01:05.843494 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.002908 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-fernet-keys\") pod \"d252c223-b923-4821-a4ff-8c183277565d\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.002989 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2klb\" (UniqueName: \"kubernetes.io/projected/d252c223-b923-4821-a4ff-8c183277565d-kube-api-access-p2klb\") pod \"d252c223-b923-4821-a4ff-8c183277565d\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.003062 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-config-data\") pod \"d252c223-b923-4821-a4ff-8c183277565d\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.003206 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-combined-ca-bundle\") pod \"d252c223-b923-4821-a4ff-8c183277565d\" (UID: \"d252c223-b923-4821-a4ff-8c183277565d\") " Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.010537 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d252c223-b923-4821-a4ff-8c183277565d-kube-api-access-p2klb" (OuterVolumeSpecName: "kube-api-access-p2klb") pod "d252c223-b923-4821-a4ff-8c183277565d" (UID: "d252c223-b923-4821-a4ff-8c183277565d"). InnerVolumeSpecName "kube-api-access-p2klb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.015371 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d252c223-b923-4821-a4ff-8c183277565d" (UID: "d252c223-b923-4821-a4ff-8c183277565d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.043124 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d252c223-b923-4821-a4ff-8c183277565d" (UID: "d252c223-b923-4821-a4ff-8c183277565d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.097548 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-config-data" (OuterVolumeSpecName: "config-data") pod "d252c223-b923-4821-a4ff-8c183277565d" (UID: "d252c223-b923-4821-a4ff-8c183277565d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.108089 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.108127 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.108138 4788 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d252c223-b923-4821-a4ff-8c183277565d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.108177 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2klb\" (UniqueName: \"kubernetes.io/projected/d252c223-b923-4821-a4ff-8c183277565d-kube-api-access-p2klb\") on node \"crc\" DevicePath \"\"" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.384019 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29335261-dvp6q" event={"ID":"d252c223-b923-4821-a4ff-8c183277565d","Type":"ContainerDied","Data":"c9f5dc4a4610c278738f5dd22f18490aa4be03aa70693c71288487ccfb4252ac"} Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.384392 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9f5dc4a4610c278738f5dd22f18490aa4be03aa70693c71288487ccfb4252ac" Oct 10 17:01:06 crc kubenswrapper[4788]: I1010 17:01:06.384114 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29335261-dvp6q" Oct 10 17:01:29 crc kubenswrapper[4788]: I1010 17:01:29.405596 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:01:29 crc kubenswrapper[4788]: I1010 17:01:29.406088 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:01:29 crc kubenswrapper[4788]: I1010 17:01:29.406131 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 17:01:29 crc kubenswrapper[4788]: I1010 17:01:29.406950 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2827f4232c96f62563197d686ef21e1ef96a8c95d1b753f0c3a009ab743fbaeb"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 17:01:29 crc kubenswrapper[4788]: I1010 17:01:29.407003 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://2827f4232c96f62563197d686ef21e1ef96a8c95d1b753f0c3a009ab743fbaeb" gracePeriod=600 Oct 10 17:01:29 crc kubenswrapper[4788]: I1010 17:01:29.597838 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="2827f4232c96f62563197d686ef21e1ef96a8c95d1b753f0c3a009ab743fbaeb" exitCode=0 Oct 10 17:01:29 crc kubenswrapper[4788]: I1010 17:01:29.598127 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"2827f4232c96f62563197d686ef21e1ef96a8c95d1b753f0c3a009ab743fbaeb"} Oct 10 17:01:29 crc kubenswrapper[4788]: I1010 17:01:29.598415 4788 scope.go:117] "RemoveContainer" containerID="b8564b3a6e5de13190d57e85ceab06b67417716d892601c029185ecfe67c2363" Oct 10 17:01:30 crc kubenswrapper[4788]: I1010 17:01:30.612781 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9"} Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.485106 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q5h8k"] Oct 10 17:03:07 crc kubenswrapper[4788]: E1010 17:03:07.486267 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d252c223-b923-4821-a4ff-8c183277565d" containerName="keystone-cron" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.486284 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d252c223-b923-4821-a4ff-8c183277565d" containerName="keystone-cron" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.486556 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d252c223-b923-4821-a4ff-8c183277565d" containerName="keystone-cron" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.488553 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.506527 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q5h8k"] Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.605951 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk72c\" (UniqueName: \"kubernetes.io/projected/b7996aa4-848b-464f-93a0-581495756d98-kube-api-access-wk72c\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.606338 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-catalog-content\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.606754 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-utilities\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.709862 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-utilities\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.710053 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk72c\" (UniqueName: \"kubernetes.io/projected/b7996aa4-848b-464f-93a0-581495756d98-kube-api-access-wk72c\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.710180 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-catalog-content\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.710469 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-utilities\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.710801 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-catalog-content\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.745420 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk72c\" (UniqueName: \"kubernetes.io/projected/b7996aa4-848b-464f-93a0-581495756d98-kube-api-access-wk72c\") pod \"redhat-operators-q5h8k\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:07 crc kubenswrapper[4788]: I1010 17:03:07.820118 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.111062 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6wb"] Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.127131 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.131442 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6wb"] Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.223573 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-catalog-content\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.223729 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-utilities\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.223910 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jv6q\" (UniqueName: \"kubernetes.io/projected/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-kube-api-access-8jv6q\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.325613 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-catalog-content\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.325713 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-utilities\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.325850 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jv6q\" (UniqueName: \"kubernetes.io/projected/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-kube-api-access-8jv6q\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.326929 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-catalog-content\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.327222 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-utilities\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.382810 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jv6q\" (UniqueName: \"kubernetes.io/projected/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-kube-api-access-8jv6q\") pod \"redhat-marketplace-jq6wb\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.403688 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q5h8k"] Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.460261 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.825676 4788 generic.go:334] "Generic (PLEG): container finished" podID="b7996aa4-848b-464f-93a0-581495756d98" containerID="7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823" exitCode=0 Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.825761 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5h8k" event={"ID":"b7996aa4-848b-464f-93a0-581495756d98","Type":"ContainerDied","Data":"7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823"} Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.826093 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5h8k" event={"ID":"b7996aa4-848b-464f-93a0-581495756d98","Type":"ContainerStarted","Data":"1130c9f0eb913e3f2569be677152012569b42adf1aec65a64ceec2e30adc8ff8"} Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.828649 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 17:03:08 crc kubenswrapper[4788]: I1010 17:03:08.967446 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6wb"] Oct 10 17:03:08 crc kubenswrapper[4788]: W1010 17:03:08.977123 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d734ed8_aa11_4e18_8d99_7195fc52b4ed.slice/crio-9e8ea204a4d0db6f489891a9917ace32f6e060a3d6109f061ea8ff23b75a763b WatchSource:0}: Error finding container 9e8ea204a4d0db6f489891a9917ace32f6e060a3d6109f061ea8ff23b75a763b: Status 404 returned error can't find the container with id 9e8ea204a4d0db6f489891a9917ace32f6e060a3d6109f061ea8ff23b75a763b Oct 10 17:03:09 crc kubenswrapper[4788]: I1010 17:03:09.862916 4788 generic.go:334] "Generic (PLEG): container finished" podID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerID="c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a" exitCode=0 Oct 10 17:03:09 crc kubenswrapper[4788]: I1010 17:03:09.863459 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6wb" event={"ID":"4d734ed8-aa11-4e18-8d99-7195fc52b4ed","Type":"ContainerDied","Data":"c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a"} Oct 10 17:03:09 crc kubenswrapper[4788]: I1010 17:03:09.863510 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6wb" event={"ID":"4d734ed8-aa11-4e18-8d99-7195fc52b4ed","Type":"ContainerStarted","Data":"9e8ea204a4d0db6f489891a9917ace32f6e060a3d6109f061ea8ff23b75a763b"} Oct 10 17:03:10 crc kubenswrapper[4788]: I1010 17:03:10.879115 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5h8k" event={"ID":"b7996aa4-848b-464f-93a0-581495756d98","Type":"ContainerStarted","Data":"42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56"} Oct 10 17:03:10 crc kubenswrapper[4788]: I1010 17:03:10.883670 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6wb" event={"ID":"4d734ed8-aa11-4e18-8d99-7195fc52b4ed","Type":"ContainerStarted","Data":"31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e"} Oct 10 17:03:11 crc kubenswrapper[4788]: I1010 17:03:11.908372 4788 generic.go:334] "Generic (PLEG): container finished" podID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerID="31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e" exitCode=0 Oct 10 17:03:11 crc kubenswrapper[4788]: I1010 17:03:11.911171 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6wb" event={"ID":"4d734ed8-aa11-4e18-8d99-7195fc52b4ed","Type":"ContainerDied","Data":"31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e"} Oct 10 17:03:12 crc kubenswrapper[4788]: I1010 17:03:12.921321 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6wb" event={"ID":"4d734ed8-aa11-4e18-8d99-7195fc52b4ed","Type":"ContainerStarted","Data":"b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c"} Oct 10 17:03:12 crc kubenswrapper[4788]: I1010 17:03:12.946787 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jq6wb" podStartSLOduration=2.238064438 podStartE2EDuration="4.946762157s" podCreationTimestamp="2025-10-10 17:03:08 +0000 UTC" firstStartedPulling="2025-10-10 17:03:09.867077027 +0000 UTC m=+8292.316792575" lastFinishedPulling="2025-10-10 17:03:12.575774746 +0000 UTC m=+8295.025490294" observedRunningTime="2025-10-10 17:03:12.943337065 +0000 UTC m=+8295.393052623" watchObservedRunningTime="2025-10-10 17:03:12.946762157 +0000 UTC m=+8295.396477725" Oct 10 17:03:13 crc kubenswrapper[4788]: I1010 17:03:13.957688 4788 generic.go:334] "Generic (PLEG): container finished" podID="b7996aa4-848b-464f-93a0-581495756d98" containerID="42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56" exitCode=0 Oct 10 17:03:13 crc kubenswrapper[4788]: I1010 17:03:13.958085 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5h8k" event={"ID":"b7996aa4-848b-464f-93a0-581495756d98","Type":"ContainerDied","Data":"42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56"} Oct 10 17:03:14 crc kubenswrapper[4788]: I1010 17:03:14.972415 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5h8k" event={"ID":"b7996aa4-848b-464f-93a0-581495756d98","Type":"ContainerStarted","Data":"b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2"} Oct 10 17:03:15 crc kubenswrapper[4788]: I1010 17:03:15.005127 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q5h8k" podStartSLOduration=2.387670949 podStartE2EDuration="8.005103575s" podCreationTimestamp="2025-10-10 17:03:07 +0000 UTC" firstStartedPulling="2025-10-10 17:03:08.828372724 +0000 UTC m=+8291.278088272" lastFinishedPulling="2025-10-10 17:03:14.44580534 +0000 UTC m=+8296.895520898" observedRunningTime="2025-10-10 17:03:14.998595532 +0000 UTC m=+8297.448311140" watchObservedRunningTime="2025-10-10 17:03:15.005103575 +0000 UTC m=+8297.454819113" Oct 10 17:03:17 crc kubenswrapper[4788]: I1010 17:03:17.820817 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:17 crc kubenswrapper[4788]: I1010 17:03:17.821118 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:18 crc kubenswrapper[4788]: I1010 17:03:18.461675 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:18 crc kubenswrapper[4788]: I1010 17:03:18.461765 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:18 crc kubenswrapper[4788]: I1010 17:03:18.514334 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:18 crc kubenswrapper[4788]: I1010 17:03:18.868054 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q5h8k" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="registry-server" probeResult="failure" output=< Oct 10 17:03:18 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 17:03:18 crc kubenswrapper[4788]: > Oct 10 17:03:19 crc kubenswrapper[4788]: I1010 17:03:19.119821 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:19 crc kubenswrapper[4788]: I1010 17:03:19.207134 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6wb"] Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.044630 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jq6wb" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerName="registry-server" containerID="cri-o://b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c" gracePeriod=2 Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.591749 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.717318 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-utilities\") pod \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.718026 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-catalog-content\") pod \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.718246 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jv6q\" (UniqueName: \"kubernetes.io/projected/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-kube-api-access-8jv6q\") pod \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\" (UID: \"4d734ed8-aa11-4e18-8d99-7195fc52b4ed\") " Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.719165 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-utilities" (OuterVolumeSpecName: "utilities") pod "4d734ed8-aa11-4e18-8d99-7195fc52b4ed" (UID: "4d734ed8-aa11-4e18-8d99-7195fc52b4ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.724502 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-kube-api-access-8jv6q" (OuterVolumeSpecName: "kube-api-access-8jv6q") pod "4d734ed8-aa11-4e18-8d99-7195fc52b4ed" (UID: "4d734ed8-aa11-4e18-8d99-7195fc52b4ed"). InnerVolumeSpecName "kube-api-access-8jv6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.738939 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d734ed8-aa11-4e18-8d99-7195fc52b4ed" (UID: "4d734ed8-aa11-4e18-8d99-7195fc52b4ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.821291 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.821327 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jv6q\" (UniqueName: \"kubernetes.io/projected/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-kube-api-access-8jv6q\") on node \"crc\" DevicePath \"\"" Oct 10 17:03:21 crc kubenswrapper[4788]: I1010 17:03:21.821341 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d734ed8-aa11-4e18-8d99-7195fc52b4ed-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.056978 4788 generic.go:334] "Generic (PLEG): container finished" podID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerID="b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c" exitCode=0 Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.057058 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6wb" event={"ID":"4d734ed8-aa11-4e18-8d99-7195fc52b4ed","Type":"ContainerDied","Data":"b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c"} Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.057106 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq6wb" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.057126 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq6wb" event={"ID":"4d734ed8-aa11-4e18-8d99-7195fc52b4ed","Type":"ContainerDied","Data":"9e8ea204a4d0db6f489891a9917ace32f6e060a3d6109f061ea8ff23b75a763b"} Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.057189 4788 scope.go:117] "RemoveContainer" containerID="b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.086301 4788 scope.go:117] "RemoveContainer" containerID="31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.106122 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6wb"] Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.116621 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq6wb"] Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.139321 4788 scope.go:117] "RemoveContainer" containerID="c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.173925 4788 scope.go:117] "RemoveContainer" containerID="b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c" Oct 10 17:03:22 crc kubenswrapper[4788]: E1010 17:03:22.174407 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c\": container with ID starting with b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c not found: ID does not exist" containerID="b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.174442 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c"} err="failed to get container status \"b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c\": rpc error: code = NotFound desc = could not find container \"b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c\": container with ID starting with b1624b9ad51ce97423dab4a7d539136777942911853faf19789a589bedb65f2c not found: ID does not exist" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.174464 4788 scope.go:117] "RemoveContainer" containerID="31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e" Oct 10 17:03:22 crc kubenswrapper[4788]: E1010 17:03:22.174827 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e\": container with ID starting with 31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e not found: ID does not exist" containerID="31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.174856 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e"} err="failed to get container status \"31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e\": rpc error: code = NotFound desc = could not find container \"31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e\": container with ID starting with 31e89ca3c2f3f0348ba17e0e7268f30ea2c4e51cc73881cd3ac70385bf04b97e not found: ID does not exist" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.174876 4788 scope.go:117] "RemoveContainer" containerID="c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a" Oct 10 17:03:22 crc kubenswrapper[4788]: E1010 17:03:22.175334 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a\": container with ID starting with c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a not found: ID does not exist" containerID="c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.175396 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a"} err="failed to get container status \"c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a\": rpc error: code = NotFound desc = could not find container \"c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a\": container with ID starting with c7ccc0033e822ff0aaebf49ece0ec55f019bd75299a461ce5bb4665b6ba2b60a not found: ID does not exist" Oct 10 17:03:22 crc kubenswrapper[4788]: I1010 17:03:22.248662 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" path="/var/lib/kubelet/pods/4d734ed8-aa11-4e18-8d99-7195fc52b4ed/volumes" Oct 10 17:03:27 crc kubenswrapper[4788]: I1010 17:03:27.904187 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:27 crc kubenswrapper[4788]: I1010 17:03:27.976943 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:28 crc kubenswrapper[4788]: I1010 17:03:28.168485 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q5h8k"] Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.182506 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q5h8k" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="registry-server" containerID="cri-o://b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2" gracePeriod=2 Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.406211 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.406723 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.698924 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.843946 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-utilities\") pod \"b7996aa4-848b-464f-93a0-581495756d98\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.844187 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-catalog-content\") pod \"b7996aa4-848b-464f-93a0-581495756d98\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.845463 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-utilities" (OuterVolumeSpecName: "utilities") pod "b7996aa4-848b-464f-93a0-581495756d98" (UID: "b7996aa4-848b-464f-93a0-581495756d98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.854428 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wk72c\" (UniqueName: \"kubernetes.io/projected/b7996aa4-848b-464f-93a0-581495756d98-kube-api-access-wk72c\") pod \"b7996aa4-848b-464f-93a0-581495756d98\" (UID: \"b7996aa4-848b-464f-93a0-581495756d98\") " Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.854940 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.867496 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7996aa4-848b-464f-93a0-581495756d98-kube-api-access-wk72c" (OuterVolumeSpecName: "kube-api-access-wk72c") pod "b7996aa4-848b-464f-93a0-581495756d98" (UID: "b7996aa4-848b-464f-93a0-581495756d98"). InnerVolumeSpecName "kube-api-access-wk72c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.932452 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7996aa4-848b-464f-93a0-581495756d98" (UID: "b7996aa4-848b-464f-93a0-581495756d98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.956864 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7996aa4-848b-464f-93a0-581495756d98-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:03:29 crc kubenswrapper[4788]: I1010 17:03:29.957080 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wk72c\" (UniqueName: \"kubernetes.io/projected/b7996aa4-848b-464f-93a0-581495756d98-kube-api-access-wk72c\") on node \"crc\" DevicePath \"\"" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.198066 4788 generic.go:334] "Generic (PLEG): container finished" podID="b7996aa4-848b-464f-93a0-581495756d98" containerID="b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2" exitCode=0 Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.198180 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5h8k" event={"ID":"b7996aa4-848b-464f-93a0-581495756d98","Type":"ContainerDied","Data":"b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2"} Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.199480 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5h8k" event={"ID":"b7996aa4-848b-464f-93a0-581495756d98","Type":"ContainerDied","Data":"1130c9f0eb913e3f2569be677152012569b42adf1aec65a64ceec2e30adc8ff8"} Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.199559 4788 scope.go:117] "RemoveContainer" containerID="b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.198236 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5h8k" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.229763 4788 scope.go:117] "RemoveContainer" containerID="42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.257387 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q5h8k"] Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.270823 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q5h8k"] Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.296151 4788 scope.go:117] "RemoveContainer" containerID="7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.341723 4788 scope.go:117] "RemoveContainer" containerID="b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2" Oct 10 17:03:30 crc kubenswrapper[4788]: E1010 17:03:30.342179 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2\": container with ID starting with b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2 not found: ID does not exist" containerID="b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.342227 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2"} err="failed to get container status \"b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2\": rpc error: code = NotFound desc = could not find container \"b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2\": container with ID starting with b753e4988af64e94f3fd518661b841b0abdfa43793f7c3bbb6ece78f6aa976c2 not found: ID does not exist" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.342248 4788 scope.go:117] "RemoveContainer" containerID="42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56" Oct 10 17:03:30 crc kubenswrapper[4788]: E1010 17:03:30.342621 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56\": container with ID starting with 42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56 not found: ID does not exist" containerID="42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.342642 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56"} err="failed to get container status \"42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56\": rpc error: code = NotFound desc = could not find container \"42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56\": container with ID starting with 42af301164085fd7b436886479b2e0415403a57e75566d65d4c5a83393078c56 not found: ID does not exist" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.342655 4788 scope.go:117] "RemoveContainer" containerID="7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823" Oct 10 17:03:30 crc kubenswrapper[4788]: E1010 17:03:30.343125 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823\": container with ID starting with 7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823 not found: ID does not exist" containerID="7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823" Oct 10 17:03:30 crc kubenswrapper[4788]: I1010 17:03:30.343175 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823"} err="failed to get container status \"7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823\": rpc error: code = NotFound desc = could not find container \"7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823\": container with ID starting with 7145c4e1d65a6c32d8cebc97d456f057897f51cd5bd413a95fc244f5cd721823 not found: ID does not exist" Oct 10 17:03:32 crc kubenswrapper[4788]: I1010 17:03:32.258338 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7996aa4-848b-464f-93a0-581495756d98" path="/var/lib/kubelet/pods/b7996aa4-848b-464f-93a0-581495756d98/volumes" Oct 10 17:03:59 crc kubenswrapper[4788]: I1010 17:03:59.406318 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:03:59 crc kubenswrapper[4788]: I1010 17:03:59.407059 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:04:16 crc kubenswrapper[4788]: I1010 17:04:16.735256 4788 generic.go:334] "Generic (PLEG): container finished" podID="e1c3f042-7a1a-4a21-accd-636a15e88010" containerID="13ea88933c752f55224fc19c564aefa885eb7fef401ccbf63d0a8b3b392e7239" exitCode=0 Oct 10 17:04:16 crc kubenswrapper[4788]: I1010 17:04:16.735334 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" event={"ID":"e1c3f042-7a1a-4a21-accd-636a15e88010","Type":"ContainerDied","Data":"13ea88933c752f55224fc19c564aefa885eb7fef401ccbf63d0a8b3b392e7239"} Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.248622 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393467 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-inventory\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393555 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-1\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393613 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-1\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393674 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-combined-ca-bundle\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393745 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-1\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393762 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-0\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393781 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-0\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393803 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t7cz\" (UniqueName: \"kubernetes.io/projected/e1c3f042-7a1a-4a21-accd-636a15e88010-kube-api-access-2t7cz\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393846 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ssh-key\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393881 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ceph\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.393903 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-0\") pod \"e1c3f042-7a1a-4a21-accd-636a15e88010\" (UID: \"e1c3f042-7a1a-4a21-accd-636a15e88010\") " Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.399686 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ceph" (OuterVolumeSpecName: "ceph") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.403064 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c3f042-7a1a-4a21-accd-636a15e88010-kube-api-access-2t7cz" (OuterVolumeSpecName: "kube-api-access-2t7cz") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "kube-api-access-2t7cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.407377 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.422759 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.428443 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.428744 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.430294 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.430973 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.437695 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.438041 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-inventory" (OuterVolumeSpecName: "inventory") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.457746 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "e1c3f042-7a1a-4a21-accd-636a15e88010" (UID: "e1c3f042-7a1a-4a21-accd-636a15e88010"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497246 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497285 4788 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497322 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497386 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497401 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497414 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497426 4788 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497438 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t7cz\" (UniqueName: \"kubernetes.io/projected/e1c3f042-7a1a-4a21-accd-636a15e88010-kube-api-access-2t7cz\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497449 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497464 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.497475 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e1c3f042-7a1a-4a21-accd-636a15e88010-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.759064 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" event={"ID":"e1c3f042-7a1a-4a21-accd-636a15e88010","Type":"ContainerDied","Data":"5015a2289475291125302c96331d84250dbe69c4f0415661604feee7afdbb2cf"} Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.759116 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5015a2289475291125302c96331d84250dbe69c4f0415661604feee7afdbb2cf" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.759189 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-cjc4m" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.864696 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-2ddfx"] Oct 10 17:04:18 crc kubenswrapper[4788]: E1010 17:04:18.865160 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerName="registry-server" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865176 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerName="registry-server" Oct 10 17:04:18 crc kubenswrapper[4788]: E1010 17:04:18.865195 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerName="extract-content" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865202 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerName="extract-content" Oct 10 17:04:18 crc kubenswrapper[4788]: E1010 17:04:18.865217 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="extract-utilities" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865224 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="extract-utilities" Oct 10 17:04:18 crc kubenswrapper[4788]: E1010 17:04:18.865237 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerName="extract-utilities" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865243 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerName="extract-utilities" Oct 10 17:04:18 crc kubenswrapper[4788]: E1010 17:04:18.865249 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c3f042-7a1a-4a21-accd-636a15e88010" containerName="nova-cell1-openstack-openstack-cell1" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865255 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c3f042-7a1a-4a21-accd-636a15e88010" containerName="nova-cell1-openstack-openstack-cell1" Oct 10 17:04:18 crc kubenswrapper[4788]: E1010 17:04:18.865274 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="extract-content" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865280 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="extract-content" Oct 10 17:04:18 crc kubenswrapper[4788]: E1010 17:04:18.865290 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="registry-server" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865296 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="registry-server" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865491 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7996aa4-848b-464f-93a0-581495756d98" containerName="registry-server" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865511 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d734ed8-aa11-4e18-8d99-7195fc52b4ed" containerName="registry-server" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.865526 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c3f042-7a1a-4a21-accd-636a15e88010" containerName="nova-cell1-openstack-openstack-cell1" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.866242 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.869205 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.869306 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.871676 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.872318 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.875282 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 17:04:18 crc kubenswrapper[4788]: I1010 17:04:18.885236 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-2ddfx"] Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.015917 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26b96\" (UniqueName: \"kubernetes.io/projected/e4a0577d-9b0e-44ae-9b49-14374f508c62-kube-api-access-26b96\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.016407 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-inventory\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.016460 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.016491 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ssh-key\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.016511 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.016534 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.016570 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceph\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.016714 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.118399 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-inventory\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.118479 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.118517 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ssh-key\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.118544 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.118579 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.118615 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceph\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.118824 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.119497 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26b96\" (UniqueName: \"kubernetes.io/projected/e4a0577d-9b0e-44ae-9b49-14374f508c62-kube-api-access-26b96\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.123627 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceph\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.124176 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ssh-key\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.124173 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.124190 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.124307 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.126279 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-inventory\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.134042 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.137924 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26b96\" (UniqueName: \"kubernetes.io/projected/e4a0577d-9b0e-44ae-9b49-14374f508c62-kube-api-access-26b96\") pod \"telemetry-openstack-openstack-cell1-2ddfx\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.187018 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:04:19 crc kubenswrapper[4788]: I1010 17:04:19.827582 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-2ddfx"] Oct 10 17:04:20 crc kubenswrapper[4788]: I1010 17:04:20.780422 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" event={"ID":"e4a0577d-9b0e-44ae-9b49-14374f508c62","Type":"ContainerStarted","Data":"1d062572087c37ff4085e6eddc0dce396de5cbe8c1f272139fbbec6eeb886b5c"} Oct 10 17:04:20 crc kubenswrapper[4788]: I1010 17:04:20.781737 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" event={"ID":"e4a0577d-9b0e-44ae-9b49-14374f508c62","Type":"ContainerStarted","Data":"45964a52df019b27e4ce35e98a2746bd31be8aee148afeacbfc4f9ebbd40f1fc"} Oct 10 17:04:20 crc kubenswrapper[4788]: I1010 17:04:20.805103 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" podStartSLOduration=2.586927931 podStartE2EDuration="2.805082198s" podCreationTimestamp="2025-10-10 17:04:18 +0000 UTC" firstStartedPulling="2025-10-10 17:04:19.828226216 +0000 UTC m=+8362.277941764" lastFinishedPulling="2025-10-10 17:04:20.046380483 +0000 UTC m=+8362.496096031" observedRunningTime="2025-10-10 17:04:20.795702049 +0000 UTC m=+8363.245417627" watchObservedRunningTime="2025-10-10 17:04:20.805082198 +0000 UTC m=+8363.254797746" Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.406726 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.407763 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.407902 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.409845 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.409967 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" gracePeriod=600 Oct 10 17:04:29 crc kubenswrapper[4788]: E1010 17:04:29.582093 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.918458 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" exitCode=0 Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.918551 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9"} Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.918660 4788 scope.go:117] "RemoveContainer" containerID="2827f4232c96f62563197d686ef21e1ef96a8c95d1b753f0c3a009ab743fbaeb" Oct 10 17:04:29 crc kubenswrapper[4788]: I1010 17:04:29.920116 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:04:29 crc kubenswrapper[4788]: E1010 17:04:29.920845 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:04:45 crc kubenswrapper[4788]: I1010 17:04:45.234332 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:04:45 crc kubenswrapper[4788]: E1010 17:04:45.235697 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:04:57 crc kubenswrapper[4788]: I1010 17:04:57.234450 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:04:57 crc kubenswrapper[4788]: E1010 17:04:57.236097 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:05:09 crc kubenswrapper[4788]: I1010 17:05:09.235082 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:05:09 crc kubenswrapper[4788]: E1010 17:05:09.235897 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:05:20 crc kubenswrapper[4788]: I1010 17:05:20.237458 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:05:20 crc kubenswrapper[4788]: E1010 17:05:20.238446 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:05:34 crc kubenswrapper[4788]: I1010 17:05:34.234272 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:05:34 crc kubenswrapper[4788]: E1010 17:05:34.235877 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:05:48 crc kubenswrapper[4788]: I1010 17:05:48.242050 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:05:48 crc kubenswrapper[4788]: E1010 17:05:48.242842 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.107763 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-84nzb"] Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.112078 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.153127 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84nzb"] Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.233102 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471efb95-ace7-444f-8786-f9eed7209eba-catalog-content\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.233280 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471efb95-ace7-444f-8786-f9eed7209eba-utilities\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.233562 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqjb\" (UniqueName: \"kubernetes.io/projected/471efb95-ace7-444f-8786-f9eed7209eba-kube-api-access-nrqjb\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.335717 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqjb\" (UniqueName: \"kubernetes.io/projected/471efb95-ace7-444f-8786-f9eed7209eba-kube-api-access-nrqjb\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.335846 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471efb95-ace7-444f-8786-f9eed7209eba-catalog-content\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.335903 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471efb95-ace7-444f-8786-f9eed7209eba-utilities\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.336312 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471efb95-ace7-444f-8786-f9eed7209eba-catalog-content\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.336477 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471efb95-ace7-444f-8786-f9eed7209eba-utilities\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.356375 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqjb\" (UniqueName: \"kubernetes.io/projected/471efb95-ace7-444f-8786-f9eed7209eba-kube-api-access-nrqjb\") pod \"certified-operators-84nzb\" (UID: \"471efb95-ace7-444f-8786-f9eed7209eba\") " pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.451363 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.947472 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84nzb"] Oct 10 17:05:57 crc kubenswrapper[4788]: I1010 17:05:57.986863 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84nzb" event={"ID":"471efb95-ace7-444f-8786-f9eed7209eba","Type":"ContainerStarted","Data":"41b03100af291b6d575ae503cd3f2c3f4354715a9df5b9b8f14469e98bac4177"} Oct 10 17:05:58 crc kubenswrapper[4788]: I1010 17:05:58.998405 4788 generic.go:334] "Generic (PLEG): container finished" podID="471efb95-ace7-444f-8786-f9eed7209eba" containerID="aa01c6ecfb5bb42290ed456a1fd8a58b3e644206d3f7bbee91cbcf7a965a12e2" exitCode=0 Oct 10 17:05:58 crc kubenswrapper[4788]: I1010 17:05:58.998483 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84nzb" event={"ID":"471efb95-ace7-444f-8786-f9eed7209eba","Type":"ContainerDied","Data":"aa01c6ecfb5bb42290ed456a1fd8a58b3e644206d3f7bbee91cbcf7a965a12e2"} Oct 10 17:05:59 crc kubenswrapper[4788]: I1010 17:05:59.234256 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:05:59 crc kubenswrapper[4788]: E1010 17:05:59.234682 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:06:04 crc kubenswrapper[4788]: I1010 17:06:04.053208 4788 generic.go:334] "Generic (PLEG): container finished" podID="471efb95-ace7-444f-8786-f9eed7209eba" containerID="dc39a15c5fbfee0b3da9fc929c772c3719dcb29eb4e704d9b711576564bb99a1" exitCode=0 Oct 10 17:06:04 crc kubenswrapper[4788]: I1010 17:06:04.053309 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84nzb" event={"ID":"471efb95-ace7-444f-8786-f9eed7209eba","Type":"ContainerDied","Data":"dc39a15c5fbfee0b3da9fc929c772c3719dcb29eb4e704d9b711576564bb99a1"} Oct 10 17:06:05 crc kubenswrapper[4788]: I1010 17:06:05.069218 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84nzb" event={"ID":"471efb95-ace7-444f-8786-f9eed7209eba","Type":"ContainerStarted","Data":"b8d99127c7dc7fbe8defab15b51706405499ded9ab759a89e27ead2d18c3790b"} Oct 10 17:06:05 crc kubenswrapper[4788]: I1010 17:06:05.099232 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-84nzb" podStartSLOduration=2.675328004 podStartE2EDuration="8.099129853s" podCreationTimestamp="2025-10-10 17:05:57 +0000 UTC" firstStartedPulling="2025-10-10 17:05:59.00148067 +0000 UTC m=+8461.451196218" lastFinishedPulling="2025-10-10 17:06:04.425282519 +0000 UTC m=+8466.874998067" observedRunningTime="2025-10-10 17:06:05.096520594 +0000 UTC m=+8467.546236142" watchObservedRunningTime="2025-10-10 17:06:05.099129853 +0000 UTC m=+8467.548845401" Oct 10 17:06:07 crc kubenswrapper[4788]: I1010 17:06:07.451892 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:06:07 crc kubenswrapper[4788]: I1010 17:06:07.452489 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:06:07 crc kubenswrapper[4788]: I1010 17:06:07.512218 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:06:13 crc kubenswrapper[4788]: I1010 17:06:13.234563 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:06:13 crc kubenswrapper[4788]: E1010 17:06:13.235740 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:06:17 crc kubenswrapper[4788]: I1010 17:06:17.544881 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-84nzb" Oct 10 17:06:17 crc kubenswrapper[4788]: I1010 17:06:17.645614 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84nzb"] Oct 10 17:06:17 crc kubenswrapper[4788]: I1010 17:06:17.703036 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6nm9r"] Oct 10 17:06:17 crc kubenswrapper[4788]: I1010 17:06:17.703503 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6nm9r" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerName="registry-server" containerID="cri-o://55ae2ee72bf11735cd34b6d37141e8eaded13fd074e197811668f0073ea3b790" gracePeriod=2 Oct 10 17:06:18 crc kubenswrapper[4788]: E1010 17:06:18.054995 4788 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod582c358b_0330_4157_bf4c_a95f9d2f30fe.slice/crio-conmon-55ae2ee72bf11735cd34b6d37141e8eaded13fd074e197811668f0073ea3b790.scope\": RecentStats: unable to find data in memory cache]" Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.266667 4788 generic.go:334] "Generic (PLEG): container finished" podID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerID="55ae2ee72bf11735cd34b6d37141e8eaded13fd074e197811668f0073ea3b790" exitCode=0 Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.266765 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nm9r" event={"ID":"582c358b-0330-4157-bf4c-a95f9d2f30fe","Type":"ContainerDied","Data":"55ae2ee72bf11735cd34b6d37141e8eaded13fd074e197811668f0073ea3b790"} Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.267228 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nm9r" event={"ID":"582c358b-0330-4157-bf4c-a95f9d2f30fe","Type":"ContainerDied","Data":"d9fe526680110540f27358664d1c50f05a5cb76c73ee044c79e8daad14dbe619"} Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.267396 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9fe526680110540f27358664d1c50f05a5cb76c73ee044c79e8daad14dbe619" Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.276463 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.392995 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfjcn\" (UniqueName: \"kubernetes.io/projected/582c358b-0330-4157-bf4c-a95f9d2f30fe-kube-api-access-nfjcn\") pod \"582c358b-0330-4157-bf4c-a95f9d2f30fe\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.393201 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-catalog-content\") pod \"582c358b-0330-4157-bf4c-a95f9d2f30fe\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.393253 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-utilities\") pod \"582c358b-0330-4157-bf4c-a95f9d2f30fe\" (UID: \"582c358b-0330-4157-bf4c-a95f9d2f30fe\") " Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.395851 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-utilities" (OuterVolumeSpecName: "utilities") pod "582c358b-0330-4157-bf4c-a95f9d2f30fe" (UID: "582c358b-0330-4157-bf4c-a95f9d2f30fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.404346 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/582c358b-0330-4157-bf4c-a95f9d2f30fe-kube-api-access-nfjcn" (OuterVolumeSpecName: "kube-api-access-nfjcn") pod "582c358b-0330-4157-bf4c-a95f9d2f30fe" (UID: "582c358b-0330-4157-bf4c-a95f9d2f30fe"). InnerVolumeSpecName "kube-api-access-nfjcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.478687 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "582c358b-0330-4157-bf4c-a95f9d2f30fe" (UID: "582c358b-0330-4157-bf4c-a95f9d2f30fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.496042 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfjcn\" (UniqueName: \"kubernetes.io/projected/582c358b-0330-4157-bf4c-a95f9d2f30fe-kube-api-access-nfjcn\") on node \"crc\" DevicePath \"\"" Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.497235 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:06:18 crc kubenswrapper[4788]: I1010 17:06:18.497328 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c358b-0330-4157-bf4c-a95f9d2f30fe-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:06:19 crc kubenswrapper[4788]: I1010 17:06:19.276691 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6nm9r" Oct 10 17:06:19 crc kubenswrapper[4788]: I1010 17:06:19.313131 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6nm9r"] Oct 10 17:06:19 crc kubenswrapper[4788]: I1010 17:06:19.318761 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6nm9r"] Oct 10 17:06:20 crc kubenswrapper[4788]: I1010 17:06:20.250341 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" path="/var/lib/kubelet/pods/582c358b-0330-4157-bf4c-a95f9d2f30fe/volumes" Oct 10 17:06:20 crc kubenswrapper[4788]: I1010 17:06:20.567818 4788 scope.go:117] "RemoveContainer" containerID="9ab9cf4dcb94528b62de1aa4cb6d4ac442122a03ffc7a038894e6740d1ac9135" Oct 10 17:06:20 crc kubenswrapper[4788]: I1010 17:06:20.596388 4788 scope.go:117] "RemoveContainer" containerID="dd0beab431b27abe49d7b4e8daf1d9a90bba99d71d65d4705da11bea0797c614" Oct 10 17:06:20 crc kubenswrapper[4788]: I1010 17:06:20.642182 4788 scope.go:117] "RemoveContainer" containerID="55ae2ee72bf11735cd34b6d37141e8eaded13fd074e197811668f0073ea3b790" Oct 10 17:06:25 crc kubenswrapper[4788]: I1010 17:06:25.234525 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:06:25 crc kubenswrapper[4788]: E1010 17:06:25.235453 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:06:39 crc kubenswrapper[4788]: I1010 17:06:39.234712 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:06:39 crc kubenswrapper[4788]: E1010 17:06:39.236745 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:06:51 crc kubenswrapper[4788]: I1010 17:06:51.235215 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:06:51 crc kubenswrapper[4788]: E1010 17:06:51.238305 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:07:02 crc kubenswrapper[4788]: I1010 17:07:02.235051 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:07:02 crc kubenswrapper[4788]: E1010 17:07:02.236034 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:07:13 crc kubenswrapper[4788]: I1010 17:07:13.235797 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:07:13 crc kubenswrapper[4788]: E1010 17:07:13.237397 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:07:25 crc kubenswrapper[4788]: I1010 17:07:25.234805 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:07:25 crc kubenswrapper[4788]: E1010 17:07:25.235736 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:07:36 crc kubenswrapper[4788]: I1010 17:07:36.234372 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:07:36 crc kubenswrapper[4788]: E1010 17:07:36.235468 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:07:51 crc kubenswrapper[4788]: I1010 17:07:51.234768 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:07:51 crc kubenswrapper[4788]: E1010 17:07:51.235803 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:08:03 crc kubenswrapper[4788]: I1010 17:08:03.234457 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:08:03 crc kubenswrapper[4788]: E1010 17:08:03.235305 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:08:15 crc kubenswrapper[4788]: I1010 17:08:15.234194 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:08:15 crc kubenswrapper[4788]: E1010 17:08:15.235071 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:08:29 crc kubenswrapper[4788]: I1010 17:08:29.234753 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:08:29 crc kubenswrapper[4788]: E1010 17:08:29.235890 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:08:30 crc kubenswrapper[4788]: I1010 17:08:30.891267 4788 generic.go:334] "Generic (PLEG): container finished" podID="e4a0577d-9b0e-44ae-9b49-14374f508c62" containerID="1d062572087c37ff4085e6eddc0dce396de5cbe8c1f272139fbbec6eeb886b5c" exitCode=0 Oct 10 17:08:30 crc kubenswrapper[4788]: I1010 17:08:30.891366 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" event={"ID":"e4a0577d-9b0e-44ae-9b49-14374f508c62","Type":"ContainerDied","Data":"1d062572087c37ff4085e6eddc0dce396de5cbe8c1f272139fbbec6eeb886b5c"} Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.389555 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.571318 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-telemetry-combined-ca-bundle\") pod \"e4a0577d-9b0e-44ae-9b49-14374f508c62\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.571430 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceph\") pod \"e4a0577d-9b0e-44ae-9b49-14374f508c62\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.571557 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-2\") pod \"e4a0577d-9b0e-44ae-9b49-14374f508c62\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.571719 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-inventory\") pod \"e4a0577d-9b0e-44ae-9b49-14374f508c62\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.572574 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-0\") pod \"e4a0577d-9b0e-44ae-9b49-14374f508c62\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.572719 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ssh-key\") pod \"e4a0577d-9b0e-44ae-9b49-14374f508c62\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.572817 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-1\") pod \"e4a0577d-9b0e-44ae-9b49-14374f508c62\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.572853 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26b96\" (UniqueName: \"kubernetes.io/projected/e4a0577d-9b0e-44ae-9b49-14374f508c62-kube-api-access-26b96\") pod \"e4a0577d-9b0e-44ae-9b49-14374f508c62\" (UID: \"e4a0577d-9b0e-44ae-9b49-14374f508c62\") " Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.578189 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4a0577d-9b0e-44ae-9b49-14374f508c62-kube-api-access-26b96" (OuterVolumeSpecName: "kube-api-access-26b96") pod "e4a0577d-9b0e-44ae-9b49-14374f508c62" (UID: "e4a0577d-9b0e-44ae-9b49-14374f508c62"). InnerVolumeSpecName "kube-api-access-26b96". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.582738 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceph" (OuterVolumeSpecName: "ceph") pod "e4a0577d-9b0e-44ae-9b49-14374f508c62" (UID: "e4a0577d-9b0e-44ae-9b49-14374f508c62"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.583120 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e4a0577d-9b0e-44ae-9b49-14374f508c62" (UID: "e4a0577d-9b0e-44ae-9b49-14374f508c62"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.603017 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-inventory" (OuterVolumeSpecName: "inventory") pod "e4a0577d-9b0e-44ae-9b49-14374f508c62" (UID: "e4a0577d-9b0e-44ae-9b49-14374f508c62"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.606632 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "e4a0577d-9b0e-44ae-9b49-14374f508c62" (UID: "e4a0577d-9b0e-44ae-9b49-14374f508c62"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.613568 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e4a0577d-9b0e-44ae-9b49-14374f508c62" (UID: "e4a0577d-9b0e-44ae-9b49-14374f508c62"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.620135 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "e4a0577d-9b0e-44ae-9b49-14374f508c62" (UID: "e4a0577d-9b0e-44ae-9b49-14374f508c62"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.627780 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "e4a0577d-9b0e-44ae-9b49-14374f508c62" (UID: "e4a0577d-9b0e-44ae-9b49-14374f508c62"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.677473 4788 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.677522 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.677536 4788 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.677579 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.677597 4788 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.677610 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.677621 4788 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e4a0577d-9b0e-44ae-9b49-14374f508c62-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.677633 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26b96\" (UniqueName: \"kubernetes.io/projected/e4a0577d-9b0e-44ae-9b49-14374f508c62-kube-api-access-26b96\") on node \"crc\" DevicePath \"\"" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.919986 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" event={"ID":"e4a0577d-9b0e-44ae-9b49-14374f508c62","Type":"ContainerDied","Data":"45964a52df019b27e4ce35e98a2746bd31be8aee148afeacbfc4f9ebbd40f1fc"} Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.920026 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45964a52df019b27e4ce35e98a2746bd31be8aee148afeacbfc4f9ebbd40f1fc" Oct 10 17:08:32 crc kubenswrapper[4788]: I1010 17:08:32.920061 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-2ddfx" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.025284 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-6xzwg"] Oct 10 17:08:33 crc kubenswrapper[4788]: E1010 17:08:33.025744 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4a0577d-9b0e-44ae-9b49-14374f508c62" containerName="telemetry-openstack-openstack-cell1" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.025765 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4a0577d-9b0e-44ae-9b49-14374f508c62" containerName="telemetry-openstack-openstack-cell1" Oct 10 17:08:33 crc kubenswrapper[4788]: E1010 17:08:33.025796 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerName="extract-utilities" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.025802 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerName="extract-utilities" Oct 10 17:08:33 crc kubenswrapper[4788]: E1010 17:08:33.025823 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerName="extract-content" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.025830 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerName="extract-content" Oct 10 17:08:33 crc kubenswrapper[4788]: E1010 17:08:33.025837 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerName="registry-server" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.025843 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerName="registry-server" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.026058 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="582c358b-0330-4157-bf4c-a95f9d2f30fe" containerName="registry-server" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.026091 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4a0577d-9b0e-44ae-9b49-14374f508c62" containerName="telemetry-openstack-openstack-cell1" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.026864 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.038306 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.038652 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.043701 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.043729 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.043831 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.068184 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-6xzwg"] Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.188018 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.188117 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-429k8\" (UniqueName: \"kubernetes.io/projected/0f131e4c-500d-4914-9b97-869a2765d81e-kube-api-access-429k8\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.188207 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.188237 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.188308 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.188483 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.290890 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.291291 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.291887 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.291953 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-429k8\" (UniqueName: \"kubernetes.io/projected/0f131e4c-500d-4914-9b97-869a2765d81e-kube-api-access-429k8\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.291982 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.292003 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.299123 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.299410 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.299538 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.312727 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.327297 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.332343 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-429k8\" (UniqueName: \"kubernetes.io/projected/0f131e4c-500d-4914-9b97-869a2765d81e-kube-api-access-429k8\") pod \"neutron-sriov-openstack-openstack-cell1-6xzwg\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.369667 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.952513 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-6xzwg"] Oct 10 17:08:33 crc kubenswrapper[4788]: I1010 17:08:33.977790 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 17:08:34 crc kubenswrapper[4788]: I1010 17:08:34.944092 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" event={"ID":"0f131e4c-500d-4914-9b97-869a2765d81e","Type":"ContainerStarted","Data":"aaab479ece8df7d5d3b5c61c0efc55177df7100de07c14e3ecd6ca82746fe6d2"} Oct 10 17:08:34 crc kubenswrapper[4788]: I1010 17:08:34.944681 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" event={"ID":"0f131e4c-500d-4914-9b97-869a2765d81e","Type":"ContainerStarted","Data":"5a8bd587609a338072282d739993fbb870762a5c2a6b3590f8c03d2ddac11457"} Oct 10 17:08:34 crc kubenswrapper[4788]: I1010 17:08:34.969853 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" podStartSLOduration=2.749848389 podStartE2EDuration="2.969812344s" podCreationTimestamp="2025-10-10 17:08:32 +0000 UTC" firstStartedPulling="2025-10-10 17:08:33.975131609 +0000 UTC m=+8616.424847147" lastFinishedPulling="2025-10-10 17:08:34.195095544 +0000 UTC m=+8616.644811102" observedRunningTime="2025-10-10 17:08:34.969799024 +0000 UTC m=+8617.419514612" watchObservedRunningTime="2025-10-10 17:08:34.969812344 +0000 UTC m=+8617.419527942" Oct 10 17:08:40 crc kubenswrapper[4788]: I1010 17:08:40.235218 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:08:40 crc kubenswrapper[4788]: E1010 17:08:40.236279 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:08:54 crc kubenswrapper[4788]: I1010 17:08:54.234604 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:08:54 crc kubenswrapper[4788]: E1010 17:08:54.235845 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:09:05 crc kubenswrapper[4788]: I1010 17:09:05.234466 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:09:05 crc kubenswrapper[4788]: E1010 17:09:05.235628 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:09:17 crc kubenswrapper[4788]: I1010 17:09:17.236289 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:09:17 crc kubenswrapper[4788]: E1010 17:09:17.236998 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:09:30 crc kubenswrapper[4788]: I1010 17:09:30.234966 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:09:30 crc kubenswrapper[4788]: I1010 17:09:30.641353 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"4851c34d54841f8de03ba2d0d779b34cab4f778c51119921848ec28c21f668e5"} Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.335204 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ht29f"] Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.339190 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.359765 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ht29f"] Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.480544 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-utilities\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.481273 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-catalog-content\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.481409 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8gxb\" (UniqueName: \"kubernetes.io/projected/00eb0186-d201-40f2-bc04-0c6f36e8c52b-kube-api-access-l8gxb\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.583546 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8gxb\" (UniqueName: \"kubernetes.io/projected/00eb0186-d201-40f2-bc04-0c6f36e8c52b-kube-api-access-l8gxb\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.583732 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-utilities\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.583896 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-catalog-content\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.584239 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-utilities\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.584302 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-catalog-content\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.611581 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8gxb\" (UniqueName: \"kubernetes.io/projected/00eb0186-d201-40f2-bc04-0c6f36e8c52b-kube-api-access-l8gxb\") pod \"community-operators-ht29f\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:36 crc kubenswrapper[4788]: I1010 17:09:36.674379 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:37 crc kubenswrapper[4788]: I1010 17:09:37.228618 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ht29f"] Oct 10 17:09:37 crc kubenswrapper[4788]: I1010 17:09:37.735640 4788 generic.go:334] "Generic (PLEG): container finished" podID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerID="0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539" exitCode=0 Oct 10 17:09:37 crc kubenswrapper[4788]: I1010 17:09:37.735722 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ht29f" event={"ID":"00eb0186-d201-40f2-bc04-0c6f36e8c52b","Type":"ContainerDied","Data":"0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539"} Oct 10 17:09:37 crc kubenswrapper[4788]: I1010 17:09:37.736114 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ht29f" event={"ID":"00eb0186-d201-40f2-bc04-0c6f36e8c52b","Type":"ContainerStarted","Data":"934f7c7ec9f932d42cf03a8b2823d302e29d6b2302aaac6cc1463c35ef88769f"} Oct 10 17:09:39 crc kubenswrapper[4788]: I1010 17:09:39.762373 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ht29f" event={"ID":"00eb0186-d201-40f2-bc04-0c6f36e8c52b","Type":"ContainerStarted","Data":"77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841"} Oct 10 17:09:40 crc kubenswrapper[4788]: I1010 17:09:40.780696 4788 generic.go:334] "Generic (PLEG): container finished" podID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerID="77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841" exitCode=0 Oct 10 17:09:40 crc kubenswrapper[4788]: I1010 17:09:40.780795 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ht29f" event={"ID":"00eb0186-d201-40f2-bc04-0c6f36e8c52b","Type":"ContainerDied","Data":"77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841"} Oct 10 17:09:41 crc kubenswrapper[4788]: I1010 17:09:41.794097 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ht29f" event={"ID":"00eb0186-d201-40f2-bc04-0c6f36e8c52b","Type":"ContainerStarted","Data":"ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d"} Oct 10 17:09:41 crc kubenswrapper[4788]: I1010 17:09:41.823722 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ht29f" podStartSLOduration=2.327367742 podStartE2EDuration="5.823696174s" podCreationTimestamp="2025-10-10 17:09:36 +0000 UTC" firstStartedPulling="2025-10-10 17:09:37.739263063 +0000 UTC m=+8680.188978611" lastFinishedPulling="2025-10-10 17:09:41.235591475 +0000 UTC m=+8683.685307043" observedRunningTime="2025-10-10 17:09:41.812000903 +0000 UTC m=+8684.261716471" watchObservedRunningTime="2025-10-10 17:09:41.823696174 +0000 UTC m=+8684.273411722" Oct 10 17:09:46 crc kubenswrapper[4788]: I1010 17:09:46.675108 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:46 crc kubenswrapper[4788]: I1010 17:09:46.675672 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:46 crc kubenswrapper[4788]: I1010 17:09:46.769057 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:46 crc kubenswrapper[4788]: I1010 17:09:46.914889 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:47 crc kubenswrapper[4788]: I1010 17:09:47.015677 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ht29f"] Oct 10 17:09:48 crc kubenswrapper[4788]: I1010 17:09:48.868897 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ht29f" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerName="registry-server" containerID="cri-o://ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d" gracePeriod=2 Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.398429 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.527536 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8gxb\" (UniqueName: \"kubernetes.io/projected/00eb0186-d201-40f2-bc04-0c6f36e8c52b-kube-api-access-l8gxb\") pod \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.528031 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-utilities\") pod \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.528081 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-catalog-content\") pod \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\" (UID: \"00eb0186-d201-40f2-bc04-0c6f36e8c52b\") " Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.529732 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-utilities" (OuterVolumeSpecName: "utilities") pod "00eb0186-d201-40f2-bc04-0c6f36e8c52b" (UID: "00eb0186-d201-40f2-bc04-0c6f36e8c52b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.534846 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00eb0186-d201-40f2-bc04-0c6f36e8c52b-kube-api-access-l8gxb" (OuterVolumeSpecName: "kube-api-access-l8gxb") pod "00eb0186-d201-40f2-bc04-0c6f36e8c52b" (UID: "00eb0186-d201-40f2-bc04-0c6f36e8c52b"). InnerVolumeSpecName "kube-api-access-l8gxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.586629 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00eb0186-d201-40f2-bc04-0c6f36e8c52b" (UID: "00eb0186-d201-40f2-bc04-0c6f36e8c52b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.631813 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.631851 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00eb0186-d201-40f2-bc04-0c6f36e8c52b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.631865 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8gxb\" (UniqueName: \"kubernetes.io/projected/00eb0186-d201-40f2-bc04-0c6f36e8c52b-kube-api-access-l8gxb\") on node \"crc\" DevicePath \"\"" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.880370 4788 generic.go:334] "Generic (PLEG): container finished" podID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerID="ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d" exitCode=0 Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.880415 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ht29f" event={"ID":"00eb0186-d201-40f2-bc04-0c6f36e8c52b","Type":"ContainerDied","Data":"ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d"} Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.880444 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ht29f" event={"ID":"00eb0186-d201-40f2-bc04-0c6f36e8c52b","Type":"ContainerDied","Data":"934f7c7ec9f932d42cf03a8b2823d302e29d6b2302aaac6cc1463c35ef88769f"} Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.880463 4788 scope.go:117] "RemoveContainer" containerID="ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.880502 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ht29f" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.908478 4788 scope.go:117] "RemoveContainer" containerID="77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.938966 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ht29f"] Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.950001 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ht29f"] Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.960503 4788 scope.go:117] "RemoveContainer" containerID="0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.994792 4788 scope.go:117] "RemoveContainer" containerID="ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d" Oct 10 17:09:49 crc kubenswrapper[4788]: E1010 17:09:49.995331 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d\": container with ID starting with ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d not found: ID does not exist" containerID="ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.995373 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d"} err="failed to get container status \"ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d\": rpc error: code = NotFound desc = could not find container \"ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d\": container with ID starting with ce90ae28a3d78b52a1592d7d9f8b53b911ef54ff55beed6abcd5139b8893618d not found: ID does not exist" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.995401 4788 scope.go:117] "RemoveContainer" containerID="77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841" Oct 10 17:09:49 crc kubenswrapper[4788]: E1010 17:09:49.995904 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841\": container with ID starting with 77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841 not found: ID does not exist" containerID="77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.995961 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841"} err="failed to get container status \"77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841\": rpc error: code = NotFound desc = could not find container \"77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841\": container with ID starting with 77e19253c98a852681cea794e496699ec093ea50c3bde87c8bad6c4b37fec841 not found: ID does not exist" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.996004 4788 scope.go:117] "RemoveContainer" containerID="0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539" Oct 10 17:09:49 crc kubenswrapper[4788]: E1010 17:09:49.996979 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539\": container with ID starting with 0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539 not found: ID does not exist" containerID="0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539" Oct 10 17:09:49 crc kubenswrapper[4788]: I1010 17:09:49.997072 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539"} err="failed to get container status \"0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539\": rpc error: code = NotFound desc = could not find container \"0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539\": container with ID starting with 0be2272c3e82d5c829730db775165a5eaef6caa265bdf76ed12404f571a8b539 not found: ID does not exist" Oct 10 17:09:50 crc kubenswrapper[4788]: I1010 17:09:50.255853 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" path="/var/lib/kubelet/pods/00eb0186-d201-40f2-bc04-0c6f36e8c52b/volumes" Oct 10 17:10:47 crc kubenswrapper[4788]: I1010 17:10:47.626086 4788 generic.go:334] "Generic (PLEG): container finished" podID="0f131e4c-500d-4914-9b97-869a2765d81e" containerID="aaab479ece8df7d5d3b5c61c0efc55177df7100de07c14e3ecd6ca82746fe6d2" exitCode=0 Oct 10 17:10:47 crc kubenswrapper[4788]: I1010 17:10:47.626230 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" event={"ID":"0f131e4c-500d-4914-9b97-869a2765d81e","Type":"ContainerDied","Data":"aaab479ece8df7d5d3b5c61c0efc55177df7100de07c14e3ecd6ca82746fe6d2"} Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.060321 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.101243 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-429k8\" (UniqueName: \"kubernetes.io/projected/0f131e4c-500d-4914-9b97-869a2765d81e-kube-api-access-429k8\") pod \"0f131e4c-500d-4914-9b97-869a2765d81e\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.101543 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ceph\") pod \"0f131e4c-500d-4914-9b97-869a2765d81e\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.101711 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-inventory\") pod \"0f131e4c-500d-4914-9b97-869a2765d81e\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.101788 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-combined-ca-bundle\") pod \"0f131e4c-500d-4914-9b97-869a2765d81e\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.101864 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-agent-neutron-config-0\") pod \"0f131e4c-500d-4914-9b97-869a2765d81e\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.101903 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ssh-key\") pod \"0f131e4c-500d-4914-9b97-869a2765d81e\" (UID: \"0f131e4c-500d-4914-9b97-869a2765d81e\") " Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.108472 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f131e4c-500d-4914-9b97-869a2765d81e-kube-api-access-429k8" (OuterVolumeSpecName: "kube-api-access-429k8") pod "0f131e4c-500d-4914-9b97-869a2765d81e" (UID: "0f131e4c-500d-4914-9b97-869a2765d81e"). InnerVolumeSpecName "kube-api-access-429k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.109626 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ceph" (OuterVolumeSpecName: "ceph") pod "0f131e4c-500d-4914-9b97-869a2765d81e" (UID: "0f131e4c-500d-4914-9b97-869a2765d81e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.112752 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "0f131e4c-500d-4914-9b97-869a2765d81e" (UID: "0f131e4c-500d-4914-9b97-869a2765d81e"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.134081 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "0f131e4c-500d-4914-9b97-869a2765d81e" (UID: "0f131e4c-500d-4914-9b97-869a2765d81e"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.137346 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-inventory" (OuterVolumeSpecName: "inventory") pod "0f131e4c-500d-4914-9b97-869a2765d81e" (UID: "0f131e4c-500d-4914-9b97-869a2765d81e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.140057 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0f131e4c-500d-4914-9b97-869a2765d81e" (UID: "0f131e4c-500d-4914-9b97-869a2765d81e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.204103 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.204164 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.204185 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-429k8\" (UniqueName: \"kubernetes.io/projected/0f131e4c-500d-4914-9b97-869a2765d81e-kube-api-access-429k8\") on node \"crc\" DevicePath \"\"" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.204201 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.204215 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.204228 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f131e4c-500d-4914-9b97-869a2765d81e-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.646424 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" event={"ID":"0f131e4c-500d-4914-9b97-869a2765d81e","Type":"ContainerDied","Data":"5a8bd587609a338072282d739993fbb870762a5c2a6b3590f8c03d2ddac11457"} Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.646726 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a8bd587609a338072282d739993fbb870762a5c2a6b3590f8c03d2ddac11457" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.646513 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-6xzwg" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.771888 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-29wxw"] Oct 10 17:10:49 crc kubenswrapper[4788]: E1010 17:10:49.772533 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f131e4c-500d-4914-9b97-869a2765d81e" containerName="neutron-sriov-openstack-openstack-cell1" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.772560 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f131e4c-500d-4914-9b97-869a2765d81e" containerName="neutron-sriov-openstack-openstack-cell1" Oct 10 17:10:49 crc kubenswrapper[4788]: E1010 17:10:49.772592 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerName="registry-server" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.772601 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerName="registry-server" Oct 10 17:10:49 crc kubenswrapper[4788]: E1010 17:10:49.772628 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerName="extract-content" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.772636 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerName="extract-content" Oct 10 17:10:49 crc kubenswrapper[4788]: E1010 17:10:49.772669 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerName="extract-utilities" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.772677 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerName="extract-utilities" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.772957 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="00eb0186-d201-40f2-bc04-0c6f36e8c52b" containerName="registry-server" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.772984 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f131e4c-500d-4914-9b97-869a2765d81e" containerName="neutron-sriov-openstack-openstack-cell1" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.774114 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.777289 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.783865 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-29wxw"] Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.785119 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.785306 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.785321 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.785430 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.917860 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrfl8\" (UniqueName: \"kubernetes.io/projected/6bc23201-2e25-46eb-a733-4585e2356f16-kube-api-access-wrfl8\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.917922 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.917966 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.918005 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.918265 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:49 crc kubenswrapper[4788]: I1010 17:10:49.918670 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.020231 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.020545 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.020689 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.020876 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.021008 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrfl8\" (UniqueName: \"kubernetes.io/projected/6bc23201-2e25-46eb-a733-4585e2356f16-kube-api-access-wrfl8\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.021120 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.024842 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.025035 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.025755 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.026217 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.027100 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.037137 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrfl8\" (UniqueName: \"kubernetes.io/projected/6bc23201-2e25-46eb-a733-4585e2356f16-kube-api-access-wrfl8\") pod \"neutron-dhcp-openstack-openstack-cell1-29wxw\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.103551 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:10:50 crc kubenswrapper[4788]: I1010 17:10:50.663360 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-29wxw"] Oct 10 17:10:51 crc kubenswrapper[4788]: I1010 17:10:51.667443 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" event={"ID":"6bc23201-2e25-46eb-a733-4585e2356f16","Type":"ContainerStarted","Data":"0057d610c4a76caa2df209931618177b6eaf0cc3c87805401148629e2116c123"} Oct 10 17:10:51 crc kubenswrapper[4788]: I1010 17:10:51.667777 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" event={"ID":"6bc23201-2e25-46eb-a733-4585e2356f16","Type":"ContainerStarted","Data":"ad2a73675205d2368b1aaa36e8adee0bfb78fe7a59141fb776ea56285d90f90b"} Oct 10 17:10:51 crc kubenswrapper[4788]: I1010 17:10:51.686715 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" podStartSLOduration=2.498972282 podStartE2EDuration="2.686688831s" podCreationTimestamp="2025-10-10 17:10:49 +0000 UTC" firstStartedPulling="2025-10-10 17:10:50.666924911 +0000 UTC m=+8753.116640459" lastFinishedPulling="2025-10-10 17:10:50.85464146 +0000 UTC m=+8753.304357008" observedRunningTime="2025-10-10 17:10:51.685269313 +0000 UTC m=+8754.134984861" watchObservedRunningTime="2025-10-10 17:10:51.686688831 +0000 UTC m=+8754.136404379" Oct 10 17:11:59 crc kubenswrapper[4788]: I1010 17:11:59.406598 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:11:59 crc kubenswrapper[4788]: I1010 17:11:59.407389 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:12:29 crc kubenswrapper[4788]: I1010 17:12:29.405918 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:12:29 crc kubenswrapper[4788]: I1010 17:12:29.406751 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:12:59 crc kubenswrapper[4788]: I1010 17:12:59.406415 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:12:59 crc kubenswrapper[4788]: I1010 17:12:59.408165 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:12:59 crc kubenswrapper[4788]: I1010 17:12:59.408317 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 17:12:59 crc kubenswrapper[4788]: I1010 17:12:59.409415 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4851c34d54841f8de03ba2d0d779b34cab4f778c51119921848ec28c21f668e5"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 17:12:59 crc kubenswrapper[4788]: I1010 17:12:59.409590 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://4851c34d54841f8de03ba2d0d779b34cab4f778c51119921848ec28c21f668e5" gracePeriod=600 Oct 10 17:13:00 crc kubenswrapper[4788]: I1010 17:13:00.263510 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="4851c34d54841f8de03ba2d0d779b34cab4f778c51119921848ec28c21f668e5" exitCode=0 Oct 10 17:13:00 crc kubenswrapper[4788]: I1010 17:13:00.263605 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"4851c34d54841f8de03ba2d0d779b34cab4f778c51119921848ec28c21f668e5"} Oct 10 17:13:00 crc kubenswrapper[4788]: I1010 17:13:00.264281 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0"} Oct 10 17:13:00 crc kubenswrapper[4788]: I1010 17:13:00.264305 4788 scope.go:117] "RemoveContainer" containerID="746e4221fb309248d08bd4b4ec772858e0969c95df84d62e670be749d17ef1b9" Oct 10 17:13:37 crc kubenswrapper[4788]: I1010 17:13:37.935640 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zzk25"] Oct 10 17:13:37 crc kubenswrapper[4788]: I1010 17:13:37.942039 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:37 crc kubenswrapper[4788]: I1010 17:13:37.951201 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzk25"] Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.042543 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-catalog-content\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.042613 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bld7\" (UniqueName: \"kubernetes.io/projected/e0f363d4-86d8-46da-804e-847822983df8-kube-api-access-5bld7\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.042637 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-utilities\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.145279 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-catalog-content\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.145331 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bld7\" (UniqueName: \"kubernetes.io/projected/e0f363d4-86d8-46da-804e-847822983df8-kube-api-access-5bld7\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.145352 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-utilities\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.146041 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-utilities\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.146047 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-catalog-content\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.167780 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bld7\" (UniqueName: \"kubernetes.io/projected/e0f363d4-86d8-46da-804e-847822983df8-kube-api-access-5bld7\") pod \"redhat-operators-zzk25\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.271429 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:38 crc kubenswrapper[4788]: I1010 17:13:38.798987 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzk25"] Oct 10 17:13:39 crc kubenswrapper[4788]: I1010 17:13:39.750127 4788 generic.go:334] "Generic (PLEG): container finished" podID="e0f363d4-86d8-46da-804e-847822983df8" containerID="8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e" exitCode=0 Oct 10 17:13:39 crc kubenswrapper[4788]: I1010 17:13:39.750231 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzk25" event={"ID":"e0f363d4-86d8-46da-804e-847822983df8","Type":"ContainerDied","Data":"8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e"} Oct 10 17:13:39 crc kubenswrapper[4788]: I1010 17:13:39.750867 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzk25" event={"ID":"e0f363d4-86d8-46da-804e-847822983df8","Type":"ContainerStarted","Data":"5bcb6ad5ebc12c65ed947827c700633ee3b1e30edb9f8cbdd4a2fd420ec95eb3"} Oct 10 17:13:39 crc kubenswrapper[4788]: I1010 17:13:39.753581 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 17:13:41 crc kubenswrapper[4788]: I1010 17:13:41.775056 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzk25" event={"ID":"e0f363d4-86d8-46da-804e-847822983df8","Type":"ContainerStarted","Data":"df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6"} Oct 10 17:13:44 crc kubenswrapper[4788]: I1010 17:13:44.811563 4788 generic.go:334] "Generic (PLEG): container finished" podID="e0f363d4-86d8-46da-804e-847822983df8" containerID="df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6" exitCode=0 Oct 10 17:13:44 crc kubenswrapper[4788]: I1010 17:13:44.811737 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzk25" event={"ID":"e0f363d4-86d8-46da-804e-847822983df8","Type":"ContainerDied","Data":"df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6"} Oct 10 17:13:45 crc kubenswrapper[4788]: I1010 17:13:45.824220 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzk25" event={"ID":"e0f363d4-86d8-46da-804e-847822983df8","Type":"ContainerStarted","Data":"dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1"} Oct 10 17:13:45 crc kubenswrapper[4788]: I1010 17:13:45.849058 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zzk25" podStartSLOduration=3.306769751 podStartE2EDuration="8.849042271s" podCreationTimestamp="2025-10-10 17:13:37 +0000 UTC" firstStartedPulling="2025-10-10 17:13:39.753358931 +0000 UTC m=+8922.203074479" lastFinishedPulling="2025-10-10 17:13:45.295631451 +0000 UTC m=+8927.745346999" observedRunningTime="2025-10-10 17:13:45.845264301 +0000 UTC m=+8928.294979849" watchObservedRunningTime="2025-10-10 17:13:45.849042271 +0000 UTC m=+8928.298757819" Oct 10 17:13:48 crc kubenswrapper[4788]: I1010 17:13:48.273242 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:48 crc kubenswrapper[4788]: I1010 17:13:48.273638 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:49 crc kubenswrapper[4788]: I1010 17:13:49.347601 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zzk25" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="registry-server" probeResult="failure" output=< Oct 10 17:13:49 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 17:13:49 crc kubenswrapper[4788]: > Oct 10 17:13:58 crc kubenswrapper[4788]: I1010 17:13:58.323800 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:58 crc kubenswrapper[4788]: I1010 17:13:58.405346 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:13:58 crc kubenswrapper[4788]: I1010 17:13:58.562618 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzk25"] Oct 10 17:13:59 crc kubenswrapper[4788]: I1010 17:13:59.993327 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zzk25" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="registry-server" containerID="cri-o://dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1" gracePeriod=2 Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.521267 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.722866 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-utilities\") pod \"e0f363d4-86d8-46da-804e-847822983df8\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.723308 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bld7\" (UniqueName: \"kubernetes.io/projected/e0f363d4-86d8-46da-804e-847822983df8-kube-api-access-5bld7\") pod \"e0f363d4-86d8-46da-804e-847822983df8\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.723405 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-catalog-content\") pod \"e0f363d4-86d8-46da-804e-847822983df8\" (UID: \"e0f363d4-86d8-46da-804e-847822983df8\") " Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.723742 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-utilities" (OuterVolumeSpecName: "utilities") pod "e0f363d4-86d8-46da-804e-847822983df8" (UID: "e0f363d4-86d8-46da-804e-847822983df8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.724285 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.730629 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f363d4-86d8-46da-804e-847822983df8-kube-api-access-5bld7" (OuterVolumeSpecName: "kube-api-access-5bld7") pod "e0f363d4-86d8-46da-804e-847822983df8" (UID: "e0f363d4-86d8-46da-804e-847822983df8"). InnerVolumeSpecName "kube-api-access-5bld7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.811946 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0f363d4-86d8-46da-804e-847822983df8" (UID: "e0f363d4-86d8-46da-804e-847822983df8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.826265 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bld7\" (UniqueName: \"kubernetes.io/projected/e0f363d4-86d8-46da-804e-847822983df8-kube-api-access-5bld7\") on node \"crc\" DevicePath \"\"" Oct 10 17:14:00 crc kubenswrapper[4788]: I1010 17:14:00.826328 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f363d4-86d8-46da-804e-847822983df8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.004534 4788 generic.go:334] "Generic (PLEG): container finished" podID="e0f363d4-86d8-46da-804e-847822983df8" containerID="dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1" exitCode=0 Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.004625 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzk25" event={"ID":"e0f363d4-86d8-46da-804e-847822983df8","Type":"ContainerDied","Data":"dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1"} Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.004661 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzk25" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.004687 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzk25" event={"ID":"e0f363d4-86d8-46da-804e-847822983df8","Type":"ContainerDied","Data":"5bcb6ad5ebc12c65ed947827c700633ee3b1e30edb9f8cbdd4a2fd420ec95eb3"} Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.004711 4788 scope.go:117] "RemoveContainer" containerID="dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.026490 4788 scope.go:117] "RemoveContainer" containerID="df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.042317 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzk25"] Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.051998 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zzk25"] Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.070662 4788 scope.go:117] "RemoveContainer" containerID="8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.097503 4788 scope.go:117] "RemoveContainer" containerID="dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1" Oct 10 17:14:01 crc kubenswrapper[4788]: E1010 17:14:01.098076 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1\": container with ID starting with dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1 not found: ID does not exist" containerID="dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.098115 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1"} err="failed to get container status \"dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1\": rpc error: code = NotFound desc = could not find container \"dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1\": container with ID starting with dccc378a10e9d7401d9e24fb343486cc8121733b9cd72cec2624ca8137f089c1 not found: ID does not exist" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.098157 4788 scope.go:117] "RemoveContainer" containerID="df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6" Oct 10 17:14:01 crc kubenswrapper[4788]: E1010 17:14:01.098530 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6\": container with ID starting with df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6 not found: ID does not exist" containerID="df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.098557 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6"} err="failed to get container status \"df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6\": rpc error: code = NotFound desc = could not find container \"df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6\": container with ID starting with df0aa512513ab7ce454f739c9851651baaa0619f9fa18601b3480e75598b16c6 not found: ID does not exist" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.098574 4788 scope.go:117] "RemoveContainer" containerID="8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e" Oct 10 17:14:01 crc kubenswrapper[4788]: E1010 17:14:01.098971 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e\": container with ID starting with 8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e not found: ID does not exist" containerID="8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e" Oct 10 17:14:01 crc kubenswrapper[4788]: I1010 17:14:01.099004 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e"} err="failed to get container status \"8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e\": rpc error: code = NotFound desc = could not find container \"8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e\": container with ID starting with 8152d4d1497bfc1e83f4c0b90fdb4a230cec3d1105e5ef5d4516847a7c2ef86e not found: ID does not exist" Oct 10 17:14:02 crc kubenswrapper[4788]: I1010 17:14:02.248011 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f363d4-86d8-46da-804e-847822983df8" path="/var/lib/kubelet/pods/e0f363d4-86d8-46da-804e-847822983df8/volumes" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.438806 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k5w6r"] Oct 10 17:14:04 crc kubenswrapper[4788]: E1010 17:14:04.440374 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="extract-content" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.440397 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="extract-content" Oct 10 17:14:04 crc kubenswrapper[4788]: E1010 17:14:04.440470 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="registry-server" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.440484 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="registry-server" Oct 10 17:14:04 crc kubenswrapper[4788]: E1010 17:14:04.440519 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="extract-utilities" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.440533 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="extract-utilities" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.441086 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f363d4-86d8-46da-804e-847822983df8" containerName="registry-server" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.444276 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.455337 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5w6r"] Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.611519 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-utilities\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.612362 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj6bk\" (UniqueName: \"kubernetes.io/projected/e7f3bcbf-773f-430b-bc07-079faf17375f-kube-api-access-jj6bk\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.612422 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-catalog-content\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.714317 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-catalog-content\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.714539 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-utilities\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.714908 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-catalog-content\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.715116 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-utilities\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.715961 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj6bk\" (UniqueName: \"kubernetes.io/projected/e7f3bcbf-773f-430b-bc07-079faf17375f-kube-api-access-jj6bk\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.760750 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj6bk\" (UniqueName: \"kubernetes.io/projected/e7f3bcbf-773f-430b-bc07-079faf17375f-kube-api-access-jj6bk\") pod \"redhat-marketplace-k5w6r\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:04 crc kubenswrapper[4788]: I1010 17:14:04.807244 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:05 crc kubenswrapper[4788]: I1010 17:14:05.288959 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5w6r"] Oct 10 17:14:05 crc kubenswrapper[4788]: W1010 17:14:05.292425 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7f3bcbf_773f_430b_bc07_079faf17375f.slice/crio-d65799fd5dc8e22236dc47ae129f9202e7fca7a07597d48ca2004271be4167fc WatchSource:0}: Error finding container d65799fd5dc8e22236dc47ae129f9202e7fca7a07597d48ca2004271be4167fc: Status 404 returned error can't find the container with id d65799fd5dc8e22236dc47ae129f9202e7fca7a07597d48ca2004271be4167fc Oct 10 17:14:06 crc kubenswrapper[4788]: I1010 17:14:06.057379 4788 generic.go:334] "Generic (PLEG): container finished" podID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerID="327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3" exitCode=0 Oct 10 17:14:06 crc kubenswrapper[4788]: I1010 17:14:06.057437 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5w6r" event={"ID":"e7f3bcbf-773f-430b-bc07-079faf17375f","Type":"ContainerDied","Data":"327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3"} Oct 10 17:14:06 crc kubenswrapper[4788]: I1010 17:14:06.057777 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5w6r" event={"ID":"e7f3bcbf-773f-430b-bc07-079faf17375f","Type":"ContainerStarted","Data":"d65799fd5dc8e22236dc47ae129f9202e7fca7a07597d48ca2004271be4167fc"} Oct 10 17:14:08 crc kubenswrapper[4788]: I1010 17:14:08.083859 4788 generic.go:334] "Generic (PLEG): container finished" podID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerID="6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785" exitCode=0 Oct 10 17:14:08 crc kubenswrapper[4788]: I1010 17:14:08.083937 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5w6r" event={"ID":"e7f3bcbf-773f-430b-bc07-079faf17375f","Type":"ContainerDied","Data":"6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785"} Oct 10 17:14:09 crc kubenswrapper[4788]: I1010 17:14:09.098013 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5w6r" event={"ID":"e7f3bcbf-773f-430b-bc07-079faf17375f","Type":"ContainerStarted","Data":"6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03"} Oct 10 17:14:09 crc kubenswrapper[4788]: I1010 17:14:09.141363 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k5w6r" podStartSLOduration=2.662672282 podStartE2EDuration="5.14133244s" podCreationTimestamp="2025-10-10 17:14:04 +0000 UTC" firstStartedPulling="2025-10-10 17:14:06.061190117 +0000 UTC m=+8948.510905705" lastFinishedPulling="2025-10-10 17:14:08.539850295 +0000 UTC m=+8950.989565863" observedRunningTime="2025-10-10 17:14:09.123101856 +0000 UTC m=+8951.572817424" watchObservedRunningTime="2025-10-10 17:14:09.14133244 +0000 UTC m=+8951.591047998" Oct 10 17:14:14 crc kubenswrapper[4788]: I1010 17:14:14.807571 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:14 crc kubenswrapper[4788]: I1010 17:14:14.808096 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:14 crc kubenswrapper[4788]: I1010 17:14:14.898486 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:15 crc kubenswrapper[4788]: I1010 17:14:15.221053 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:15 crc kubenswrapper[4788]: I1010 17:14:15.264076 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5w6r"] Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.187009 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k5w6r" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerName="registry-server" containerID="cri-o://6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03" gracePeriod=2 Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.714546 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.893270 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-catalog-content\") pod \"e7f3bcbf-773f-430b-bc07-079faf17375f\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.893310 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-utilities\") pod \"e7f3bcbf-773f-430b-bc07-079faf17375f\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.893343 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj6bk\" (UniqueName: \"kubernetes.io/projected/e7f3bcbf-773f-430b-bc07-079faf17375f-kube-api-access-jj6bk\") pod \"e7f3bcbf-773f-430b-bc07-079faf17375f\" (UID: \"e7f3bcbf-773f-430b-bc07-079faf17375f\") " Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.894163 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-utilities" (OuterVolumeSpecName: "utilities") pod "e7f3bcbf-773f-430b-bc07-079faf17375f" (UID: "e7f3bcbf-773f-430b-bc07-079faf17375f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.900425 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f3bcbf-773f-430b-bc07-079faf17375f-kube-api-access-jj6bk" (OuterVolumeSpecName: "kube-api-access-jj6bk") pod "e7f3bcbf-773f-430b-bc07-079faf17375f" (UID: "e7f3bcbf-773f-430b-bc07-079faf17375f"). InnerVolumeSpecName "kube-api-access-jj6bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.915239 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7f3bcbf-773f-430b-bc07-079faf17375f" (UID: "e7f3bcbf-773f-430b-bc07-079faf17375f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.996059 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.996102 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7f3bcbf-773f-430b-bc07-079faf17375f-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:14:17 crc kubenswrapper[4788]: I1010 17:14:17.996156 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj6bk\" (UniqueName: \"kubernetes.io/projected/e7f3bcbf-773f-430b-bc07-079faf17375f-kube-api-access-jj6bk\") on node \"crc\" DevicePath \"\"" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.208981 4788 generic.go:334] "Generic (PLEG): container finished" podID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerID="6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03" exitCode=0 Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.209054 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5w6r" event={"ID":"e7f3bcbf-773f-430b-bc07-079faf17375f","Type":"ContainerDied","Data":"6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03"} Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.209118 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5w6r" event={"ID":"e7f3bcbf-773f-430b-bc07-079faf17375f","Type":"ContainerDied","Data":"d65799fd5dc8e22236dc47ae129f9202e7fca7a07597d48ca2004271be4167fc"} Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.209061 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k5w6r" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.209202 4788 scope.go:117] "RemoveContainer" containerID="6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.240631 4788 scope.go:117] "RemoveContainer" containerID="6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.271850 4788 scope.go:117] "RemoveContainer" containerID="327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.281457 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5w6r"] Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.290552 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5w6r"] Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.353699 4788 scope.go:117] "RemoveContainer" containerID="6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03" Oct 10 17:14:18 crc kubenswrapper[4788]: E1010 17:14:18.354340 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03\": container with ID starting with 6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03 not found: ID does not exist" containerID="6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.354385 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03"} err="failed to get container status \"6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03\": rpc error: code = NotFound desc = could not find container \"6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03\": container with ID starting with 6f9d47cc6819734841e7914a4cbf66d3059975b7249597dd0ebcb50222b58b03 not found: ID does not exist" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.354415 4788 scope.go:117] "RemoveContainer" containerID="6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785" Oct 10 17:14:18 crc kubenswrapper[4788]: E1010 17:14:18.354948 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785\": container with ID starting with 6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785 not found: ID does not exist" containerID="6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.354972 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785"} err="failed to get container status \"6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785\": rpc error: code = NotFound desc = could not find container \"6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785\": container with ID starting with 6318dfb4e80c0eb62a240e6b31fb84e855d38116a6797a44e7aab668d098b785 not found: ID does not exist" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.354987 4788 scope.go:117] "RemoveContainer" containerID="327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3" Oct 10 17:14:18 crc kubenswrapper[4788]: E1010 17:14:18.355334 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3\": container with ID starting with 327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3 not found: ID does not exist" containerID="327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3" Oct 10 17:14:18 crc kubenswrapper[4788]: I1010 17:14:18.355354 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3"} err="failed to get container status \"327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3\": rpc error: code = NotFound desc = could not find container \"327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3\": container with ID starting with 327f96ae9b178e00fdf1a17f034be84a8f723365035392cef6cfaf02fe84ebd3 not found: ID does not exist" Oct 10 17:14:20 crc kubenswrapper[4788]: I1010 17:14:20.270855 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" path="/var/lib/kubelet/pods/e7f3bcbf-773f-430b-bc07-079faf17375f/volumes" Oct 10 17:14:59 crc kubenswrapper[4788]: I1010 17:14:59.406248 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:14:59 crc kubenswrapper[4788]: I1010 17:14:59.407223 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.180754 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm"] Oct 10 17:15:00 crc kubenswrapper[4788]: E1010 17:15:00.181540 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerName="registry-server" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.181660 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerName="registry-server" Oct 10 17:15:00 crc kubenswrapper[4788]: E1010 17:15:00.181699 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerName="extract-utilities" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.181707 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerName="extract-utilities" Oct 10 17:15:00 crc kubenswrapper[4788]: E1010 17:15:00.181741 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerName="extract-content" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.181750 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerName="extract-content" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.181997 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f3bcbf-773f-430b-bc07-079faf17375f" containerName="registry-server" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.183134 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.186770 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.187188 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.207534 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm"] Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.329155 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b5fb73a-138e-48d6-913e-ecd699274358-secret-volume\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.329228 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bhxx\" (UniqueName: \"kubernetes.io/projected/0b5fb73a-138e-48d6-913e-ecd699274358-kube-api-access-9bhxx\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.329255 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b5fb73a-138e-48d6-913e-ecd699274358-config-volume\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.431371 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b5fb73a-138e-48d6-913e-ecd699274358-secret-volume\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.431452 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bhxx\" (UniqueName: \"kubernetes.io/projected/0b5fb73a-138e-48d6-913e-ecd699274358-kube-api-access-9bhxx\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.431490 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b5fb73a-138e-48d6-913e-ecd699274358-config-volume\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.432496 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b5fb73a-138e-48d6-913e-ecd699274358-config-volume\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.439572 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b5fb73a-138e-48d6-913e-ecd699274358-secret-volume\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.451631 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bhxx\" (UniqueName: \"kubernetes.io/projected/0b5fb73a-138e-48d6-913e-ecd699274358-kube-api-access-9bhxx\") pod \"collect-profiles-29335275-dtgbm\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:00 crc kubenswrapper[4788]: I1010 17:15:00.513103 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:01 crc kubenswrapper[4788]: I1010 17:15:01.046522 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm"] Oct 10 17:15:01 crc kubenswrapper[4788]: I1010 17:15:01.757221 4788 generic.go:334] "Generic (PLEG): container finished" podID="0b5fb73a-138e-48d6-913e-ecd699274358" containerID="d9460577af3e47e42cab030c9dc1acdc9bd58e8534d577d1259e7f439415987c" exitCode=0 Oct 10 17:15:01 crc kubenswrapper[4788]: I1010 17:15:01.757339 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" event={"ID":"0b5fb73a-138e-48d6-913e-ecd699274358","Type":"ContainerDied","Data":"d9460577af3e47e42cab030c9dc1acdc9bd58e8534d577d1259e7f439415987c"} Oct 10 17:15:01 crc kubenswrapper[4788]: I1010 17:15:01.757747 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" event={"ID":"0b5fb73a-138e-48d6-913e-ecd699274358","Type":"ContainerStarted","Data":"f3df65f7ad924c88e53af59d1bbc02612adffe13e69baa8d733ce1754dc972e2"} Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.206420 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.310123 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b5fb73a-138e-48d6-913e-ecd699274358-secret-volume\") pod \"0b5fb73a-138e-48d6-913e-ecd699274358\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.310514 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b5fb73a-138e-48d6-913e-ecd699274358-config-volume\") pod \"0b5fb73a-138e-48d6-913e-ecd699274358\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.311613 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b5fb73a-138e-48d6-913e-ecd699274358-config-volume" (OuterVolumeSpecName: "config-volume") pod "0b5fb73a-138e-48d6-913e-ecd699274358" (UID: "0b5fb73a-138e-48d6-913e-ecd699274358"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.311898 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bhxx\" (UniqueName: \"kubernetes.io/projected/0b5fb73a-138e-48d6-913e-ecd699274358-kube-api-access-9bhxx\") pod \"0b5fb73a-138e-48d6-913e-ecd699274358\" (UID: \"0b5fb73a-138e-48d6-913e-ecd699274358\") " Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.313893 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b5fb73a-138e-48d6-913e-ecd699274358-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.320349 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b5fb73a-138e-48d6-913e-ecd699274358-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0b5fb73a-138e-48d6-913e-ecd699274358" (UID: "0b5fb73a-138e-48d6-913e-ecd699274358"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.333302 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b5fb73a-138e-48d6-913e-ecd699274358-kube-api-access-9bhxx" (OuterVolumeSpecName: "kube-api-access-9bhxx") pod "0b5fb73a-138e-48d6-913e-ecd699274358" (UID: "0b5fb73a-138e-48d6-913e-ecd699274358"). InnerVolumeSpecName "kube-api-access-9bhxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.416519 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bhxx\" (UniqueName: \"kubernetes.io/projected/0b5fb73a-138e-48d6-913e-ecd699274358-kube-api-access-9bhxx\") on node \"crc\" DevicePath \"\"" Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.416551 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b5fb73a-138e-48d6-913e-ecd699274358-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.790890 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" event={"ID":"0b5fb73a-138e-48d6-913e-ecd699274358","Type":"ContainerDied","Data":"f3df65f7ad924c88e53af59d1bbc02612adffe13e69baa8d733ce1754dc972e2"} Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.790942 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3df65f7ad924c88e53af59d1bbc02612adffe13e69baa8d733ce1754dc972e2" Oct 10 17:15:03 crc kubenswrapper[4788]: I1010 17:15:03.791202 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335275-dtgbm" Oct 10 17:15:04 crc kubenswrapper[4788]: I1010 17:15:04.330301 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72"] Oct 10 17:15:04 crc kubenswrapper[4788]: I1010 17:15:04.353315 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335230-wcr72"] Oct 10 17:15:06 crc kubenswrapper[4788]: I1010 17:15:06.256612 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b448e2-395e-4e6b-9aa6-f029920da644" path="/var/lib/kubelet/pods/59b448e2-395e-4e6b-9aa6-f029920da644/volumes" Oct 10 17:15:19 crc kubenswrapper[4788]: I1010 17:15:19.025250 4788 trace.go:236] Trace[341647487]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-glv4r" (10-Oct-2025 17:15:17.864) (total time: 1160ms): Oct 10 17:15:19 crc kubenswrapper[4788]: Trace[341647487]: [1.160621976s] [1.160621976s] END Oct 10 17:15:20 crc kubenswrapper[4788]: I1010 17:15:20.954164 4788 scope.go:117] "RemoveContainer" containerID="16ef81a728b43af8397b8c84c8be6a4fb55918f928dd006e91bb1705dbed929f" Oct 10 17:15:29 crc kubenswrapper[4788]: I1010 17:15:29.406302 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:15:29 crc kubenswrapper[4788]: I1010 17:15:29.407131 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:15:59 crc kubenswrapper[4788]: I1010 17:15:59.406340 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:15:59 crc kubenswrapper[4788]: I1010 17:15:59.407123 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:15:59 crc kubenswrapper[4788]: I1010 17:15:59.407200 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 17:15:59 crc kubenswrapper[4788]: I1010 17:15:59.408281 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 17:15:59 crc kubenswrapper[4788]: I1010 17:15:59.408338 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" gracePeriod=600 Oct 10 17:15:59 crc kubenswrapper[4788]: E1010 17:15:59.548499 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:16:00 crc kubenswrapper[4788]: I1010 17:16:00.531420 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" exitCode=0 Oct 10 17:16:00 crc kubenswrapper[4788]: I1010 17:16:00.531546 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0"} Oct 10 17:16:00 crc kubenswrapper[4788]: I1010 17:16:00.531930 4788 scope.go:117] "RemoveContainer" containerID="4851c34d54841f8de03ba2d0d779b34cab4f778c51119921848ec28c21f668e5" Oct 10 17:16:00 crc kubenswrapper[4788]: I1010 17:16:00.533503 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:16:00 crc kubenswrapper[4788]: E1010 17:16:00.536767 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:16:13 crc kubenswrapper[4788]: I1010 17:16:13.234008 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:16:13 crc kubenswrapper[4788]: E1010 17:16:13.236291 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:16:28 crc kubenswrapper[4788]: I1010 17:16:28.249303 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:16:28 crc kubenswrapper[4788]: E1010 17:16:28.251974 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.234607 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:16:41 crc kubenswrapper[4788]: E1010 17:16:41.235555 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.494855 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v6nlz"] Oct 10 17:16:41 crc kubenswrapper[4788]: E1010 17:16:41.495917 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b5fb73a-138e-48d6-913e-ecd699274358" containerName="collect-profiles" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.495941 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b5fb73a-138e-48d6-913e-ecd699274358" containerName="collect-profiles" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.496254 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b5fb73a-138e-48d6-913e-ecd699274358" containerName="collect-profiles" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.499061 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.523509 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v6nlz"] Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.593157 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-catalog-content\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.593265 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srz5t\" (UniqueName: \"kubernetes.io/projected/22876de5-5118-47f9-b7a3-dccae6d3cc5c-kube-api-access-srz5t\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.593299 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-utilities\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.695942 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-catalog-content\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.696074 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srz5t\" (UniqueName: \"kubernetes.io/projected/22876de5-5118-47f9-b7a3-dccae6d3cc5c-kube-api-access-srz5t\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.696116 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-utilities\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.696647 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-utilities\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.696699 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-catalog-content\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.722907 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srz5t\" (UniqueName: \"kubernetes.io/projected/22876de5-5118-47f9-b7a3-dccae6d3cc5c-kube-api-access-srz5t\") pod \"certified-operators-v6nlz\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:41 crc kubenswrapper[4788]: I1010 17:16:41.873197 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:42 crc kubenswrapper[4788]: I1010 17:16:42.425400 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v6nlz"] Oct 10 17:16:43 crc kubenswrapper[4788]: I1010 17:16:43.057513 4788 generic.go:334] "Generic (PLEG): container finished" podID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerID="47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974" exitCode=0 Oct 10 17:16:43 crc kubenswrapper[4788]: I1010 17:16:43.057567 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6nlz" event={"ID":"22876de5-5118-47f9-b7a3-dccae6d3cc5c","Type":"ContainerDied","Data":"47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974"} Oct 10 17:16:43 crc kubenswrapper[4788]: I1010 17:16:43.057598 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6nlz" event={"ID":"22876de5-5118-47f9-b7a3-dccae6d3cc5c","Type":"ContainerStarted","Data":"c18b9d7e427f445ca95193799bc52ce17ef651bccb066fbb1b248e2ae12508ea"} Oct 10 17:16:46 crc kubenswrapper[4788]: I1010 17:16:46.094790 4788 generic.go:334] "Generic (PLEG): container finished" podID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerID="19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6" exitCode=0 Oct 10 17:16:46 crc kubenswrapper[4788]: I1010 17:16:46.095571 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6nlz" event={"ID":"22876de5-5118-47f9-b7a3-dccae6d3cc5c","Type":"ContainerDied","Data":"19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6"} Oct 10 17:16:48 crc kubenswrapper[4788]: I1010 17:16:48.127506 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6nlz" event={"ID":"22876de5-5118-47f9-b7a3-dccae6d3cc5c","Type":"ContainerStarted","Data":"c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1"} Oct 10 17:16:48 crc kubenswrapper[4788]: I1010 17:16:48.157723 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v6nlz" podStartSLOduration=3.6456530689999997 podStartE2EDuration="7.157704294s" podCreationTimestamp="2025-10-10 17:16:41 +0000 UTC" firstStartedPulling="2025-10-10 17:16:43.061535102 +0000 UTC m=+9105.511250690" lastFinishedPulling="2025-10-10 17:16:46.573586367 +0000 UTC m=+9109.023301915" observedRunningTime="2025-10-10 17:16:48.148291726 +0000 UTC m=+9110.598007274" watchObservedRunningTime="2025-10-10 17:16:48.157704294 +0000 UTC m=+9110.607419842" Oct 10 17:16:51 crc kubenswrapper[4788]: I1010 17:16:51.874299 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:51 crc kubenswrapper[4788]: I1010 17:16:51.874971 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:51 crc kubenswrapper[4788]: I1010 17:16:51.951956 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:52 crc kubenswrapper[4788]: I1010 17:16:52.224568 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:52 crc kubenswrapper[4788]: I1010 17:16:52.282880 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v6nlz"] Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.192076 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v6nlz" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerName="registry-server" containerID="cri-o://c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1" gracePeriod=2 Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.234263 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:16:54 crc kubenswrapper[4788]: E1010 17:16:54.235042 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.677386 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.768141 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-utilities\") pod \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.768270 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-catalog-content\") pod \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.768496 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srz5t\" (UniqueName: \"kubernetes.io/projected/22876de5-5118-47f9-b7a3-dccae6d3cc5c-kube-api-access-srz5t\") pod \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\" (UID: \"22876de5-5118-47f9-b7a3-dccae6d3cc5c\") " Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.769810 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-utilities" (OuterVolumeSpecName: "utilities") pod "22876de5-5118-47f9-b7a3-dccae6d3cc5c" (UID: "22876de5-5118-47f9-b7a3-dccae6d3cc5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.778059 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22876de5-5118-47f9-b7a3-dccae6d3cc5c-kube-api-access-srz5t" (OuterVolumeSpecName: "kube-api-access-srz5t") pod "22876de5-5118-47f9-b7a3-dccae6d3cc5c" (UID: "22876de5-5118-47f9-b7a3-dccae6d3cc5c"). InnerVolumeSpecName "kube-api-access-srz5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.824116 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22876de5-5118-47f9-b7a3-dccae6d3cc5c" (UID: "22876de5-5118-47f9-b7a3-dccae6d3cc5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.872075 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srz5t\" (UniqueName: \"kubernetes.io/projected/22876de5-5118-47f9-b7a3-dccae6d3cc5c-kube-api-access-srz5t\") on node \"crc\" DevicePath \"\"" Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.872116 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:16:54 crc kubenswrapper[4788]: I1010 17:16:54.872131 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22876de5-5118-47f9-b7a3-dccae6d3cc5c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.204759 4788 generic.go:334] "Generic (PLEG): container finished" podID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerID="c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1" exitCode=0 Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.204800 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6nlz" event={"ID":"22876de5-5118-47f9-b7a3-dccae6d3cc5c","Type":"ContainerDied","Data":"c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1"} Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.204826 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v6nlz" event={"ID":"22876de5-5118-47f9-b7a3-dccae6d3cc5c","Type":"ContainerDied","Data":"c18b9d7e427f445ca95193799bc52ce17ef651bccb066fbb1b248e2ae12508ea"} Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.204841 4788 scope.go:117] "RemoveContainer" containerID="c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.204855 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v6nlz" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.234199 4788 scope.go:117] "RemoveContainer" containerID="19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.241828 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v6nlz"] Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.257271 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v6nlz"] Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.273322 4788 scope.go:117] "RemoveContainer" containerID="47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.333132 4788 scope.go:117] "RemoveContainer" containerID="c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1" Oct 10 17:16:55 crc kubenswrapper[4788]: E1010 17:16:55.333696 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1\": container with ID starting with c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1 not found: ID does not exist" containerID="c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.333787 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1"} err="failed to get container status \"c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1\": rpc error: code = NotFound desc = could not find container \"c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1\": container with ID starting with c0cbab285c8dd86a276f82b4b8ea045521b11577b26b02906346abecd06b3ed1 not found: ID does not exist" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.333862 4788 scope.go:117] "RemoveContainer" containerID="19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6" Oct 10 17:16:55 crc kubenswrapper[4788]: E1010 17:16:55.334300 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6\": container with ID starting with 19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6 not found: ID does not exist" containerID="19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.334393 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6"} err="failed to get container status \"19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6\": rpc error: code = NotFound desc = could not find container \"19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6\": container with ID starting with 19cfd2e4068148ce76e15559c8c66c9f828528a0d2e09ec29df3be75703810d6 not found: ID does not exist" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.334466 4788 scope.go:117] "RemoveContainer" containerID="47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974" Oct 10 17:16:55 crc kubenswrapper[4788]: E1010 17:16:55.334798 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974\": container with ID starting with 47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974 not found: ID does not exist" containerID="47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974" Oct 10 17:16:55 crc kubenswrapper[4788]: I1010 17:16:55.334899 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974"} err="failed to get container status \"47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974\": rpc error: code = NotFound desc = could not find container \"47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974\": container with ID starting with 47e1d93cd5a7bb7198a565ae403a64c9e06eb16d980d1ffad92ca99fb08b2974 not found: ID does not exist" Oct 10 17:16:56 crc kubenswrapper[4788]: I1010 17:16:56.252183 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" path="/var/lib/kubelet/pods/22876de5-5118-47f9-b7a3-dccae6d3cc5c/volumes" Oct 10 17:17:05 crc kubenswrapper[4788]: I1010 17:17:05.235197 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:17:05 crc kubenswrapper[4788]: E1010 17:17:05.238434 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:17:20 crc kubenswrapper[4788]: I1010 17:17:20.234495 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:17:20 crc kubenswrapper[4788]: E1010 17:17:20.235593 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:17:32 crc kubenswrapper[4788]: I1010 17:17:32.235065 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:17:32 crc kubenswrapper[4788]: E1010 17:17:32.235989 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:17:43 crc kubenswrapper[4788]: I1010 17:17:43.234419 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:17:43 crc kubenswrapper[4788]: E1010 17:17:43.235352 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:17:57 crc kubenswrapper[4788]: I1010 17:17:57.235372 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:17:57 crc kubenswrapper[4788]: E1010 17:17:57.236755 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:18:08 crc kubenswrapper[4788]: I1010 17:18:08.251412 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:18:08 crc kubenswrapper[4788]: E1010 17:18:08.252702 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:18:19 crc kubenswrapper[4788]: I1010 17:18:19.234855 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:18:19 crc kubenswrapper[4788]: E1010 17:18:19.237403 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:18:31 crc kubenswrapper[4788]: I1010 17:18:31.234916 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:18:31 crc kubenswrapper[4788]: E1010 17:18:31.236326 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:18:45 crc kubenswrapper[4788]: I1010 17:18:45.234411 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:18:45 crc kubenswrapper[4788]: E1010 17:18:45.235540 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:18:59 crc kubenswrapper[4788]: I1010 17:18:59.236333 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:18:59 crc kubenswrapper[4788]: E1010 17:18:59.239895 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:19:13 crc kubenswrapper[4788]: I1010 17:19:13.235184 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:19:13 crc kubenswrapper[4788]: E1010 17:19:13.236645 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:19:25 crc kubenswrapper[4788]: I1010 17:19:25.234094 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:19:25 crc kubenswrapper[4788]: E1010 17:19:25.235108 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:19:28 crc kubenswrapper[4788]: I1010 17:19:28.212309 4788 generic.go:334] "Generic (PLEG): container finished" podID="6bc23201-2e25-46eb-a733-4585e2356f16" containerID="0057d610c4a76caa2df209931618177b6eaf0cc3c87805401148629e2116c123" exitCode=0 Oct 10 17:19:28 crc kubenswrapper[4788]: I1010 17:19:28.212392 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" event={"ID":"6bc23201-2e25-46eb-a733-4585e2356f16","Type":"ContainerDied","Data":"0057d610c4a76caa2df209931618177b6eaf0cc3c87805401148629e2116c123"} Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.230616 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.247804 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.272090 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-29wxw" event={"ID":"6bc23201-2e25-46eb-a733-4585e2356f16","Type":"ContainerDied","Data":"ad2a73675205d2368b1aaa36e8adee0bfb78fe7a59141fb776ea56285d90f90b"} Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.272201 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad2a73675205d2368b1aaa36e8adee0bfb78fe7a59141fb776ea56285d90f90b" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.296396 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-agent-neutron-config-0\") pod \"6bc23201-2e25-46eb-a733-4585e2356f16\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.296569 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ssh-key\") pod \"6bc23201-2e25-46eb-a733-4585e2356f16\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.296656 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrfl8\" (UniqueName: \"kubernetes.io/projected/6bc23201-2e25-46eb-a733-4585e2356f16-kube-api-access-wrfl8\") pod \"6bc23201-2e25-46eb-a733-4585e2356f16\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.296777 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ceph\") pod \"6bc23201-2e25-46eb-a733-4585e2356f16\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.296810 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-inventory\") pod \"6bc23201-2e25-46eb-a733-4585e2356f16\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.296918 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-combined-ca-bundle\") pod \"6bc23201-2e25-46eb-a733-4585e2356f16\" (UID: \"6bc23201-2e25-46eb-a733-4585e2356f16\") " Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.411419 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ceph" (OuterVolumeSpecName: "ceph") pod "6bc23201-2e25-46eb-a733-4585e2356f16" (UID: "6bc23201-2e25-46eb-a733-4585e2356f16"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.413694 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bc23201-2e25-46eb-a733-4585e2356f16-kube-api-access-wrfl8" (OuterVolumeSpecName: "kube-api-access-wrfl8") pod "6bc23201-2e25-46eb-a733-4585e2356f16" (UID: "6bc23201-2e25-46eb-a733-4585e2356f16"). InnerVolumeSpecName "kube-api-access-wrfl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.414553 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "6bc23201-2e25-46eb-a733-4585e2356f16" (UID: "6bc23201-2e25-46eb-a733-4585e2356f16"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.434409 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "6bc23201-2e25-46eb-a733-4585e2356f16" (UID: "6bc23201-2e25-46eb-a733-4585e2356f16"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.436372 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-inventory" (OuterVolumeSpecName: "inventory") pod "6bc23201-2e25-46eb-a733-4585e2356f16" (UID: "6bc23201-2e25-46eb-a733-4585e2356f16"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.444969 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6bc23201-2e25-46eb-a733-4585e2356f16" (UID: "6bc23201-2e25-46eb-a733-4585e2356f16"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.521370 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.521417 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.521430 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrfl8\" (UniqueName: \"kubernetes.io/projected/6bc23201-2e25-46eb-a733-4585e2356f16-kube-api-access-wrfl8\") on node \"crc\" DevicePath \"\"" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.521450 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.521460 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 17:19:30 crc kubenswrapper[4788]: I1010 17:19:30.521471 4788 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc23201-2e25-46eb-a733-4585e2356f16-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:19:38 crc kubenswrapper[4788]: I1010 17:19:38.244036 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:19:38 crc kubenswrapper[4788]: E1010 17:19:38.244839 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:19:52 crc kubenswrapper[4788]: I1010 17:19:52.234827 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:19:52 crc kubenswrapper[4788]: E1010 17:19:52.235961 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:20:00 crc kubenswrapper[4788]: I1010 17:20:00.225514 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 17:20:00 crc kubenswrapper[4788]: I1010 17:20:00.226472 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="f757adf6-3865-45a1-80ba-239090629e2b" containerName="nova-cell0-conductor-conductor" containerID="cri-o://1cc7b83df429ca373bc48fe1d50a2e0e0360637ec494af9dda855dc6bda4e0f3" gracePeriod=30 Oct 10 17:20:00 crc kubenswrapper[4788]: I1010 17:20:00.272244 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 17:20:00 crc kubenswrapper[4788]: I1010 17:20:00.272441 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="79194ba7-bc68-4165-8ab1-a6956df339cd" containerName="nova-cell1-conductor-conductor" containerID="cri-o://0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b" gracePeriod=30 Oct 10 17:20:01 crc kubenswrapper[4788]: I1010 17:20:01.837825 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 17:20:01 crc kubenswrapper[4788]: I1010 17:20:01.838718 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-log" containerID="cri-o://a7d39af77ab3d3a330b37ba43345d98cd0ff2c271758721891e58d7630c041a1" gracePeriod=30 Oct 10 17:20:01 crc kubenswrapper[4788]: I1010 17:20:01.838817 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-api" containerID="cri-o://4dda8c383b199b04672af35a66a42aa3033507d666c020e2dc7f9af35e09bda9" gracePeriod=30 Oct 10 17:20:01 crc kubenswrapper[4788]: I1010 17:20:01.857953 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 17:20:01 crc kubenswrapper[4788]: I1010 17:20:01.858217 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" containerName="nova-scheduler-scheduler" containerID="cri-o://9f762d634176aaaa4210b783680d5baf3c7c6137f69bc900ce4efa3acb49dca4" gracePeriod=30 Oct 10 17:20:01 crc kubenswrapper[4788]: I1010 17:20:01.873134 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 17:20:01 crc kubenswrapper[4788]: I1010 17:20:01.874532 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-log" containerID="cri-o://f6bbec90a1956f4b6b68625c173efca3d4de9f6019ffb56152a6cd2b1cd11cba" gracePeriod=30 Oct 10 17:20:01 crc kubenswrapper[4788]: I1010 17:20:01.875065 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-metadata" containerID="cri-o://b1f6730c0bfab5eb494145f52bca89a2c0059f624a0cfe64c7b1cf4ee89b4b39" gracePeriod=30 Oct 10 17:20:02 crc kubenswrapper[4788]: I1010 17:20:02.649451 4788 generic.go:334] "Generic (PLEG): container finished" podID="3178184c-97f1-4dfb-aea9-d969ac506208" containerID="f6bbec90a1956f4b6b68625c173efca3d4de9f6019ffb56152a6cd2b1cd11cba" exitCode=143 Oct 10 17:20:02 crc kubenswrapper[4788]: I1010 17:20:02.649742 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3178184c-97f1-4dfb-aea9-d969ac506208","Type":"ContainerDied","Data":"f6bbec90a1956f4b6b68625c173efca3d4de9f6019ffb56152a6cd2b1cd11cba"} Oct 10 17:20:02 crc kubenswrapper[4788]: I1010 17:20:02.654052 4788 generic.go:334] "Generic (PLEG): container finished" podID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerID="a7d39af77ab3d3a330b37ba43345d98cd0ff2c271758721891e58d7630c041a1" exitCode=143 Oct 10 17:20:02 crc kubenswrapper[4788]: I1010 17:20:02.654698 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcc4fc46-31f4-45d1-9255-d3f5496b6840","Type":"ContainerDied","Data":"a7d39af77ab3d3a330b37ba43345d98cd0ff2c271758721891e58d7630c041a1"} Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.671516 4788 generic.go:334] "Generic (PLEG): container finished" podID="d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" containerID="9f762d634176aaaa4210b783680d5baf3c7c6137f69bc900ce4efa3acb49dca4" exitCode=0 Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.671800 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31","Type":"ContainerDied","Data":"9f762d634176aaaa4210b783680d5baf3c7c6137f69bc900ce4efa3acb49dca4"} Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.672113 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31","Type":"ContainerDied","Data":"26f454ba5fa2fa37f8eade1832c321e8798d905466e6da1fc974eeff277df928"} Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.672162 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26f454ba5fa2fa37f8eade1832c321e8798d905466e6da1fc974eeff277df928" Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.702609 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.807158 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mbtp\" (UniqueName: \"kubernetes.io/projected/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-kube-api-access-8mbtp\") pod \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.807218 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-config-data\") pod \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.807284 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-combined-ca-bundle\") pod \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\" (UID: \"d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31\") " Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.826124 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-kube-api-access-8mbtp" (OuterVolumeSpecName: "kube-api-access-8mbtp") pod "d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" (UID: "d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31"). InnerVolumeSpecName "kube-api-access-8mbtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.852131 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-config-data" (OuterVolumeSpecName: "config-data") pod "d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" (UID: "d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.857289 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" (UID: "d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.912895 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mbtp\" (UniqueName: \"kubernetes.io/projected/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-kube-api-access-8mbtp\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.913523 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:03 crc kubenswrapper[4788]: I1010 17:20:03.913542 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.234720 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:20:04 crc kubenswrapper[4788]: E1010 17:20:04.235235 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.683985 4788 generic.go:334] "Generic (PLEG): container finished" podID="f757adf6-3865-45a1-80ba-239090629e2b" containerID="1cc7b83df429ca373bc48fe1d50a2e0e0360637ec494af9dda855dc6bda4e0f3" exitCode=0 Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.684473 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.685703 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f757adf6-3865-45a1-80ba-239090629e2b","Type":"ContainerDied","Data":"1cc7b83df429ca373bc48fe1d50a2e0e0360637ec494af9dda855dc6bda4e0f3"} Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.685751 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f757adf6-3865-45a1-80ba-239090629e2b","Type":"ContainerDied","Data":"afdf0c70421d58da3e20c29507b17570ffdc6d58a9cc0560eaa204fe596155e0"} Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.685762 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afdf0c70421d58da3e20c29507b17570ffdc6d58a9cc0560eaa204fe596155e0" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.794553 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.812721 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.827280 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.831519 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9phv2\" (UniqueName: \"kubernetes.io/projected/f757adf6-3865-45a1-80ba-239090629e2b-kube-api-access-9phv2\") pod \"f757adf6-3865-45a1-80ba-239090629e2b\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.831683 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-config-data\") pod \"f757adf6-3865-45a1-80ba-239090629e2b\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.831793 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-combined-ca-bundle\") pod \"f757adf6-3865-45a1-80ba-239090629e2b\" (UID: \"f757adf6-3865-45a1-80ba-239090629e2b\") " Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.839234 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 17:20:04 crc kubenswrapper[4788]: E1010 17:20:04.841382 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f757adf6-3865-45a1-80ba-239090629e2b" containerName="nova-cell0-conductor-conductor" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841406 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="f757adf6-3865-45a1-80ba-239090629e2b" containerName="nova-cell0-conductor-conductor" Oct 10 17:20:04 crc kubenswrapper[4788]: E1010 17:20:04.841445 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" containerName="nova-scheduler-scheduler" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841454 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" containerName="nova-scheduler-scheduler" Oct 10 17:20:04 crc kubenswrapper[4788]: E1010 17:20:04.841481 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerName="extract-content" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841500 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerName="extract-content" Oct 10 17:20:04 crc kubenswrapper[4788]: E1010 17:20:04.841513 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerName="extract-utilities" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841521 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerName="extract-utilities" Oct 10 17:20:04 crc kubenswrapper[4788]: E1010 17:20:04.841536 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerName="registry-server" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841542 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerName="registry-server" Oct 10 17:20:04 crc kubenswrapper[4788]: E1010 17:20:04.841557 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bc23201-2e25-46eb-a733-4585e2356f16" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841566 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bc23201-2e25-46eb-a733-4585e2356f16" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841780 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bc23201-2e25-46eb-a733-4585e2356f16" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841802 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="22876de5-5118-47f9-b7a3-dccae6d3cc5c" containerName="registry-server" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841815 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="f757adf6-3865-45a1-80ba-239090629e2b" containerName="nova-cell0-conductor-conductor" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.841832 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" containerName="nova-scheduler-scheduler" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.842747 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.844220 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f757adf6-3865-45a1-80ba-239090629e2b-kube-api-access-9phv2" (OuterVolumeSpecName: "kube-api-access-9phv2") pod "f757adf6-3865-45a1-80ba-239090629e2b" (UID: "f757adf6-3865-45a1-80ba-239090629e2b"). InnerVolumeSpecName "kube-api-access-9phv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.848466 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.876235 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-config-data" (OuterVolumeSpecName: "config-data") pod "f757adf6-3865-45a1-80ba-239090629e2b" (UID: "f757adf6-3865-45a1-80ba-239090629e2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.879225 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.886041 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f757adf6-3865-45a1-80ba-239090629e2b" (UID: "f757adf6-3865-45a1-80ba-239090629e2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.934630 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c967\" (UniqueName: \"kubernetes.io/projected/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-kube-api-access-8c967\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.934916 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.934966 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-config-data\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.935047 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.935059 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f757adf6-3865-45a1-80ba-239090629e2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:04 crc kubenswrapper[4788]: I1010 17:20:04.935069 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9phv2\" (UniqueName: \"kubernetes.io/projected/f757adf6-3865-45a1-80ba-239090629e2b-kube-api-access-9phv2\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.038714 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.038814 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-config-data\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.039000 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c967\" (UniqueName: \"kubernetes.io/projected/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-kube-api-access-8c967\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.043339 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.054842 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-config-data\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.057994 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c967\" (UniqueName: \"kubernetes.io/projected/6b5b6539-1514-46d5-8bfe-bfc92da4c9a0-kube-api-access-8c967\") pod \"nova-scheduler-0\" (UID: \"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0\") " pod="openstack/nova-scheduler-0" Oct 10 17:20:05 crc kubenswrapper[4788]: E1010 17:20:05.094876 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b is running failed: container process not found" containerID="0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 17:20:05 crc kubenswrapper[4788]: E1010 17:20:05.095683 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b is running failed: container process not found" containerID="0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 17:20:05 crc kubenswrapper[4788]: E1010 17:20:05.096075 4788 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b is running failed: container process not found" containerID="0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 10 17:20:05 crc kubenswrapper[4788]: E1010 17:20:05.096223 4788 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="79194ba7-bc68-4165-8ab1-a6956df339cd" containerName="nova-cell1-conductor-conductor" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.122110 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.141768 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-config-data\") pod \"79194ba7-bc68-4165-8ab1-a6956df339cd\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.141817 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-combined-ca-bundle\") pod \"79194ba7-bc68-4165-8ab1-a6956df339cd\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.141847 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmwtl\" (UniqueName: \"kubernetes.io/projected/79194ba7-bc68-4165-8ab1-a6956df339cd-kube-api-access-fmwtl\") pod \"79194ba7-bc68-4165-8ab1-a6956df339cd\" (UID: \"79194ba7-bc68-4165-8ab1-a6956df339cd\") " Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.147381 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79194ba7-bc68-4165-8ab1-a6956df339cd-kube-api-access-fmwtl" (OuterVolumeSpecName: "kube-api-access-fmwtl") pod "79194ba7-bc68-4165-8ab1-a6956df339cd" (UID: "79194ba7-bc68-4165-8ab1-a6956df339cd"). InnerVolumeSpecName "kube-api-access-fmwtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.172915 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-config-data" (OuterVolumeSpecName: "config-data") pod "79194ba7-bc68-4165-8ab1-a6956df339cd" (UID: "79194ba7-bc68-4165-8ab1-a6956df339cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.180432 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79194ba7-bc68-4165-8ab1-a6956df339cd" (UID: "79194ba7-bc68-4165-8ab1-a6956df339cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.223293 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.245529 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.245562 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79194ba7-bc68-4165-8ab1-a6956df339cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.245574 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmwtl\" (UniqueName: \"kubernetes.io/projected/79194ba7-bc68-4165-8ab1-a6956df339cd-kube-api-access-fmwtl\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.712407 4788 generic.go:334] "Generic (PLEG): container finished" podID="3178184c-97f1-4dfb-aea9-d969ac506208" containerID="b1f6730c0bfab5eb494145f52bca89a2c0059f624a0cfe64c7b1cf4ee89b4b39" exitCode=0 Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.712446 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3178184c-97f1-4dfb-aea9-d969ac506208","Type":"ContainerDied","Data":"b1f6730c0bfab5eb494145f52bca89a2c0059f624a0cfe64c7b1cf4ee89b4b39"} Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.715046 4788 generic.go:334] "Generic (PLEG): container finished" podID="79194ba7-bc68-4165-8ab1-a6956df339cd" containerID="0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b" exitCode=0 Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.715163 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.715180 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.715157 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"79194ba7-bc68-4165-8ab1-a6956df339cd","Type":"ContainerDied","Data":"0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b"} Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.715291 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"79194ba7-bc68-4165-8ab1-a6956df339cd","Type":"ContainerDied","Data":"409417680e7d64ef1b187210be125e9ea4d4b05a8ed19babc3db458fcbb4bd19"} Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.715312 4788 scope.go:117] "RemoveContainer" containerID="0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.722155 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.764926 4788 scope.go:117] "RemoveContainer" containerID="0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b" Oct 10 17:20:05 crc kubenswrapper[4788]: E1010 17:20:05.766024 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b\": container with ID starting with 0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b not found: ID does not exist" containerID="0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.766082 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b"} err="failed to get container status \"0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b\": rpc error: code = NotFound desc = could not find container \"0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b\": container with ID starting with 0eab3e183a36e58bd12981b7e9fa55f59504cf8a8e79fe17958c4d71bafe4f7b not found: ID does not exist" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.803827 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.881354 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.907252 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 17:20:05 crc kubenswrapper[4788]: E1010 17:20:05.908186 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79194ba7-bc68-4165-8ab1-a6956df339cd" containerName="nova-cell1-conductor-conductor" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.908213 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="79194ba7-bc68-4165-8ab1-a6956df339cd" containerName="nova-cell1-conductor-conductor" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.908728 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="79194ba7-bc68-4165-8ab1-a6956df339cd" containerName="nova-cell1-conductor-conductor" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.910046 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.912459 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.934771 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.961319 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.985671 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 17:20:05 crc kubenswrapper[4788]: I1010 17:20:05.997384 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.005514 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-config-data\") pod \"3178184c-97f1-4dfb-aea9-d969ac506208\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.005603 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3178184c-97f1-4dfb-aea9-d969ac506208-logs\") pod \"3178184c-97f1-4dfb-aea9-d969ac506208\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.005678 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-combined-ca-bundle\") pod \"3178184c-97f1-4dfb-aea9-d969ac506208\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.005756 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn8v2\" (UniqueName: \"kubernetes.io/projected/3178184c-97f1-4dfb-aea9-d969ac506208-kube-api-access-zn8v2\") pod \"3178184c-97f1-4dfb-aea9-d969ac506208\" (UID: \"3178184c-97f1-4dfb-aea9-d969ac506208\") " Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.007762 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3178184c-97f1-4dfb-aea9-d969ac506208-logs" (OuterVolumeSpecName: "logs") pod "3178184c-97f1-4dfb-aea9-d969ac506208" (UID: "3178184c-97f1-4dfb-aea9-d969ac506208"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.008615 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj5q9\" (UniqueName: \"kubernetes.io/projected/370a716b-2910-4149-9722-fb383dab7e4c-kube-api-access-nj5q9\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.009422 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/370a716b-2910-4149-9722-fb383dab7e4c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.009491 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370a716b-2910-4149-9722-fb383dab7e4c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.009635 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3178184c-97f1-4dfb-aea9-d969ac506208-logs\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.012806 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3178184c-97f1-4dfb-aea9-d969ac506208-kube-api-access-zn8v2" (OuterVolumeSpecName: "kube-api-access-zn8v2") pod "3178184c-97f1-4dfb-aea9-d969ac506208" (UID: "3178184c-97f1-4dfb-aea9-d969ac506208"). InnerVolumeSpecName "kube-api-access-zn8v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.014015 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 17:20:06 crc kubenswrapper[4788]: E1010 17:20:06.014509 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-metadata" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.014528 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-metadata" Oct 10 17:20:06 crc kubenswrapper[4788]: E1010 17:20:06.014557 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-log" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.014565 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-log" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.014828 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-metadata" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.014873 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" containerName="nova-metadata-log" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.015770 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.018581 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.029632 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.078882 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3178184c-97f1-4dfb-aea9-d969ac506208" (UID: "3178184c-97f1-4dfb-aea9-d969ac506208"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.086525 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-config-data" (OuterVolumeSpecName: "config-data") pod "3178184c-97f1-4dfb-aea9-d969ac506208" (UID: "3178184c-97f1-4dfb-aea9-d969ac506208"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.112680 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhnhn\" (UniqueName: \"kubernetes.io/projected/d37aad02-67ff-4f8f-af27-55d1acaed2e7-kube-api-access-fhnhn\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.112830 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37aad02-67ff-4f8f-af27-55d1acaed2e7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.113006 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37aad02-67ff-4f8f-af27-55d1acaed2e7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.113117 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/370a716b-2910-4149-9722-fb383dab7e4c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.113192 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370a716b-2910-4149-9722-fb383dab7e4c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.113292 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj5q9\" (UniqueName: \"kubernetes.io/projected/370a716b-2910-4149-9722-fb383dab7e4c-kube-api-access-nj5q9\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.113623 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.113636 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3178184c-97f1-4dfb-aea9-d969ac506208-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.113649 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn8v2\" (UniqueName: \"kubernetes.io/projected/3178184c-97f1-4dfb-aea9-d969ac506208-kube-api-access-zn8v2\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.118050 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/370a716b-2910-4149-9722-fb383dab7e4c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.119446 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/370a716b-2910-4149-9722-fb383dab7e4c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.135062 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj5q9\" (UniqueName: \"kubernetes.io/projected/370a716b-2910-4149-9722-fb383dab7e4c-kube-api-access-nj5q9\") pod \"nova-cell0-conductor-0\" (UID: \"370a716b-2910-4149-9722-fb383dab7e4c\") " pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.215084 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhnhn\" (UniqueName: \"kubernetes.io/projected/d37aad02-67ff-4f8f-af27-55d1acaed2e7-kube-api-access-fhnhn\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.215160 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37aad02-67ff-4f8f-af27-55d1acaed2e7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.215208 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37aad02-67ff-4f8f-af27-55d1acaed2e7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.220281 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37aad02-67ff-4f8f-af27-55d1acaed2e7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.222548 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37aad02-67ff-4f8f-af27-55d1acaed2e7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.233108 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhnhn\" (UniqueName: \"kubernetes.io/projected/d37aad02-67ff-4f8f-af27-55d1acaed2e7-kube-api-access-fhnhn\") pod \"nova-cell1-conductor-0\" (UID: \"d37aad02-67ff-4f8f-af27-55d1acaed2e7\") " pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.252517 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79194ba7-bc68-4165-8ab1-a6956df339cd" path="/var/lib/kubelet/pods/79194ba7-bc68-4165-8ab1-a6956df339cd/volumes" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.254052 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31" path="/var/lib/kubelet/pods/d7a31b3f-add4-43c5-b7fa-1f8c82fc5c31/volumes" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.254585 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f757adf6-3865-45a1-80ba-239090629e2b" path="/var/lib/kubelet/pods/f757adf6-3865-45a1-80ba-239090629e2b/volumes" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.280854 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.350028 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.739547 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0","Type":"ContainerStarted","Data":"562893eb54ec28c9be2c23bf933a54b927bcf2f73728de1285903c3b33505715"} Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.739997 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b5b6539-1514-46d5-8bfe-bfc92da4c9a0","Type":"ContainerStarted","Data":"c1db2e6dd37f7c2811fa6f01baf9d564d70d00c18f573e386d0b42934fdf0188"} Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.743169 4788 generic.go:334] "Generic (PLEG): container finished" podID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerID="4dda8c383b199b04672af35a66a42aa3033507d666c020e2dc7f9af35e09bda9" exitCode=0 Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.743244 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcc4fc46-31f4-45d1-9255-d3f5496b6840","Type":"ContainerDied","Data":"4dda8c383b199b04672af35a66a42aa3033507d666c020e2dc7f9af35e09bda9"} Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.745633 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3178184c-97f1-4dfb-aea9-d969ac506208","Type":"ContainerDied","Data":"aa073bac6f334aa5eb006434de3790a057039394ca7bebe1184666c66a73e4c6"} Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.745672 4788 scope.go:117] "RemoveContainer" containerID="b1f6730c0bfab5eb494145f52bca89a2c0059f624a0cfe64c7b1cf4ee89b4b39" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.745774 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.804700 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.804668863 podStartE2EDuration="2.804668863s" podCreationTimestamp="2025-10-10 17:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 17:20:06.763612802 +0000 UTC m=+9309.213328360" watchObservedRunningTime="2025-10-10 17:20:06.804668863 +0000 UTC m=+9309.254384411" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.811838 4788 scope.go:117] "RemoveContainer" containerID="f6bbec90a1956f4b6b68625c173efca3d4de9f6019ffb56152a6cd2b1cd11cba" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.844050 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.862458 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.877337 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.879333 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.892439 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.894905 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.907809 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.941956 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ed4b86-3570-4e57-a9f8-44e17a81beda-logs\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.942319 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js7mx\" (UniqueName: \"kubernetes.io/projected/19ed4b86-3570-4e57-a9f8-44e17a81beda-kube-api-access-js7mx\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.942404 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ed4b86-3570-4e57-a9f8-44e17a81beda-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.942512 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ed4b86-3570-4e57-a9f8-44e17a81beda-config-data\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:06 crc kubenswrapper[4788]: I1010 17:20:06.986788 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.043852 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwjcj\" (UniqueName: \"kubernetes.io/projected/fcc4fc46-31f4-45d1-9255-d3f5496b6840-kube-api-access-wwjcj\") pod \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.044003 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-config-data\") pod \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.044233 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc4fc46-31f4-45d1-9255-d3f5496b6840-logs\") pod \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.044268 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-combined-ca-bundle\") pod \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\" (UID: \"fcc4fc46-31f4-45d1-9255-d3f5496b6840\") " Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.044697 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ed4b86-3570-4e57-a9f8-44e17a81beda-logs\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.044809 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js7mx\" (UniqueName: \"kubernetes.io/projected/19ed4b86-3570-4e57-a9f8-44e17a81beda-kube-api-access-js7mx\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.044834 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ed4b86-3570-4e57-a9f8-44e17a81beda-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.044874 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ed4b86-3570-4e57-a9f8-44e17a81beda-config-data\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.045822 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcc4fc46-31f4-45d1-9255-d3f5496b6840-logs" (OuterVolumeSpecName: "logs") pod "fcc4fc46-31f4-45d1-9255-d3f5496b6840" (UID: "fcc4fc46-31f4-45d1-9255-d3f5496b6840"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.046322 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19ed4b86-3570-4e57-a9f8-44e17a81beda-logs\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.051635 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19ed4b86-3570-4e57-a9f8-44e17a81beda-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.053660 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcc4fc46-31f4-45d1-9255-d3f5496b6840-kube-api-access-wwjcj" (OuterVolumeSpecName: "kube-api-access-wwjcj") pod "fcc4fc46-31f4-45d1-9255-d3f5496b6840" (UID: "fcc4fc46-31f4-45d1-9255-d3f5496b6840"). InnerVolumeSpecName "kube-api-access-wwjcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.054093 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19ed4b86-3570-4e57-a9f8-44e17a81beda-config-data\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.073845 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js7mx\" (UniqueName: \"kubernetes.io/projected/19ed4b86-3570-4e57-a9f8-44e17a81beda-kube-api-access-js7mx\") pod \"nova-metadata-0\" (UID: \"19ed4b86-3570-4e57-a9f8-44e17a81beda\") " pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.081373 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-config-data" (OuterVolumeSpecName: "config-data") pod "fcc4fc46-31f4-45d1-9255-d3f5496b6840" (UID: "fcc4fc46-31f4-45d1-9255-d3f5496b6840"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.085352 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.117012 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcc4fc46-31f4-45d1-9255-d3f5496b6840" (UID: "fcc4fc46-31f4-45d1-9255-d3f5496b6840"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.147784 4788 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc4fc46-31f4-45d1-9255-d3f5496b6840-logs\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.147842 4788 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.147857 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwjcj\" (UniqueName: \"kubernetes.io/projected/fcc4fc46-31f4-45d1-9255-d3f5496b6840-kube-api-access-wwjcj\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.147867 4788 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc4fc46-31f4-45d1-9255-d3f5496b6840-config-data\") on node \"crc\" DevicePath \"\"" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.211965 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.759076 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fcc4fc46-31f4-45d1-9255-d3f5496b6840","Type":"ContainerDied","Data":"dab4d8f4c0e015878d78c52ac537e48c9bdd37a6f2ee43e8df387637c10877d8"} Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.760258 4788 scope.go:117] "RemoveContainer" containerID="4dda8c383b199b04672af35a66a42aa3033507d666c020e2dc7f9af35e09bda9" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.759157 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.761071 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d37aad02-67ff-4f8f-af27-55d1acaed2e7","Type":"ContainerStarted","Data":"0f989db1de1641b5235eba3d69124e3e42a12524fe578720865d7e2f68009161"} Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.765434 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"370a716b-2910-4149-9722-fb383dab7e4c","Type":"ContainerStarted","Data":"25fea81da80e3a1b8b89a579bbf77869e455637f409ec11954a865674161d5e0"} Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.765478 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"370a716b-2910-4149-9722-fb383dab7e4c","Type":"ContainerStarted","Data":"939563ca7e6e616acc88ed2aa4e9bbb80c6a277869b5be9cc3b9c20d1bce43e2"} Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.766938 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:07 crc kubenswrapper[4788]: I1010 17:20:07.787561 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.787539921 podStartE2EDuration="2.787539921s" podCreationTimestamp="2025-10-10 17:20:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 17:20:07.78180255 +0000 UTC m=+9310.231518098" watchObservedRunningTime="2025-10-10 17:20:07.787539921 +0000 UTC m=+9310.237255469" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.253517 4788 scope.go:117] "RemoveContainer" containerID="a7d39af77ab3d3a330b37ba43345d98cd0ff2c271758721891e58d7630c041a1" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.316587 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3178184c-97f1-4dfb-aea9-d969ac506208" path="/var/lib/kubelet/pods/3178184c-97f1-4dfb-aea9-d969ac506208/volumes" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.325572 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.350444 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.361766 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 10 17:20:08 crc kubenswrapper[4788]: E1010 17:20:08.362324 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-api" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.362344 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-api" Oct 10 17:20:08 crc kubenswrapper[4788]: E1010 17:20:08.362368 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-log" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.362375 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-log" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.362612 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-log" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.362643 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" containerName="nova-api-api" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.364202 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.366883 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.374997 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.410558 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2029803d-4cea-46d6-af6c-7a0b17a4733c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.410838 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2029803d-4cea-46d6-af6c-7a0b17a4733c-logs\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.411010 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm42z\" (UniqueName: \"kubernetes.io/projected/2029803d-4cea-46d6-af6c-7a0b17a4733c-kube-api-access-dm42z\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.411055 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2029803d-4cea-46d6-af6c-7a0b17a4733c-config-data\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.514041 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2029803d-4cea-46d6-af6c-7a0b17a4733c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.514164 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2029803d-4cea-46d6-af6c-7a0b17a4733c-logs\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.514221 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm42z\" (UniqueName: \"kubernetes.io/projected/2029803d-4cea-46d6-af6c-7a0b17a4733c-kube-api-access-dm42z\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.514243 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2029803d-4cea-46d6-af6c-7a0b17a4733c-config-data\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.515487 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2029803d-4cea-46d6-af6c-7a0b17a4733c-logs\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.521962 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2029803d-4cea-46d6-af6c-7a0b17a4733c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.522274 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2029803d-4cea-46d6-af6c-7a0b17a4733c-config-data\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.532042 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm42z\" (UniqueName: \"kubernetes.io/projected/2029803d-4cea-46d6-af6c-7a0b17a4733c-kube-api-access-dm42z\") pod \"nova-api-0\" (UID: \"2029803d-4cea-46d6-af6c-7a0b17a4733c\") " pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.588520 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 10 17:20:08 crc kubenswrapper[4788]: W1010 17:20:08.594930 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19ed4b86_3570_4e57_a9f8_44e17a81beda.slice/crio-ee648e598f42e9aa40dfabc1d8ad76579fb610460f3fe1df8e5a49d254524ac0 WatchSource:0}: Error finding container ee648e598f42e9aa40dfabc1d8ad76579fb610460f3fe1df8e5a49d254524ac0: Status 404 returned error can't find the container with id ee648e598f42e9aa40dfabc1d8ad76579fb610460f3fe1df8e5a49d254524ac0 Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.685693 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.795812 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d37aad02-67ff-4f8f-af27-55d1acaed2e7","Type":"ContainerStarted","Data":"95d1c2afcc827c578fd480cb5e5d3cdbf654c6238e36bec91fd479489878fd4b"} Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.798413 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.808455 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19ed4b86-3570-4e57-a9f8-44e17a81beda","Type":"ContainerStarted","Data":"ee648e598f42e9aa40dfabc1d8ad76579fb610460f3fe1df8e5a49d254524ac0"} Oct 10 17:20:08 crc kubenswrapper[4788]: I1010 17:20:08.820464 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.820444665 podStartE2EDuration="3.820444665s" podCreationTimestamp="2025-10-10 17:20:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 17:20:08.818188296 +0000 UTC m=+9311.267903864" watchObservedRunningTime="2025-10-10 17:20:08.820444665 +0000 UTC m=+9311.270160223" Oct 10 17:20:09 crc kubenswrapper[4788]: I1010 17:20:09.324321 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 10 17:20:09 crc kubenswrapper[4788]: W1010 17:20:09.327258 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2029803d_4cea_46d6_af6c_7a0b17a4733c.slice/crio-7a24a21431b25abbafcd4c16708c29e2fdde693170a15aa5934c8c360d92ab10 WatchSource:0}: Error finding container 7a24a21431b25abbafcd4c16708c29e2fdde693170a15aa5934c8c360d92ab10: Status 404 returned error can't find the container with id 7a24a21431b25abbafcd4c16708c29e2fdde693170a15aa5934c8c360d92ab10 Oct 10 17:20:09 crc kubenswrapper[4788]: I1010 17:20:09.823940 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19ed4b86-3570-4e57-a9f8-44e17a81beda","Type":"ContainerStarted","Data":"faba19248570fc2fab411f5b003da236ebde6fb4d501110b33a947231910f726"} Oct 10 17:20:09 crc kubenswrapper[4788]: I1010 17:20:09.824426 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19ed4b86-3570-4e57-a9f8-44e17a81beda","Type":"ContainerStarted","Data":"0f6260122a8d6441ba5644c57003c895d3e06f35b34914a331afc158c230a128"} Oct 10 17:20:09 crc kubenswrapper[4788]: I1010 17:20:09.830200 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2029803d-4cea-46d6-af6c-7a0b17a4733c","Type":"ContainerStarted","Data":"be56d9684f5569600bf671792309631fa06605cd5ea9383d052842e9dacac4c2"} Oct 10 17:20:09 crc kubenswrapper[4788]: I1010 17:20:09.830261 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2029803d-4cea-46d6-af6c-7a0b17a4733c","Type":"ContainerStarted","Data":"443ce917f023bd3c2c9b58dae0e0c6d5d91953e5132579e072e5f7e24425a322"} Oct 10 17:20:09 crc kubenswrapper[4788]: I1010 17:20:09.830275 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2029803d-4cea-46d6-af6c-7a0b17a4733c","Type":"ContainerStarted","Data":"7a24a21431b25abbafcd4c16708c29e2fdde693170a15aa5934c8c360d92ab10"} Oct 10 17:20:09 crc kubenswrapper[4788]: I1010 17:20:09.870903 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.8708680709999999 podStartE2EDuration="1.870868071s" podCreationTimestamp="2025-10-10 17:20:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 17:20:09.865993672 +0000 UTC m=+9312.315709220" watchObservedRunningTime="2025-10-10 17:20:09.870868071 +0000 UTC m=+9312.320583619" Oct 10 17:20:09 crc kubenswrapper[4788]: I1010 17:20:09.873615 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.8736022930000003 podStartE2EDuration="3.873602293s" podCreationTimestamp="2025-10-10 17:20:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-10 17:20:09.84766943 +0000 UTC m=+9312.297384988" watchObservedRunningTime="2025-10-10 17:20:09.873602293 +0000 UTC m=+9312.323317841" Oct 10 17:20:10 crc kubenswrapper[4788]: I1010 17:20:10.225434 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 10 17:20:10 crc kubenswrapper[4788]: I1010 17:20:10.252300 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcc4fc46-31f4-45d1-9255-d3f5496b6840" path="/var/lib/kubelet/pods/fcc4fc46-31f4-45d1-9255-d3f5496b6840/volumes" Oct 10 17:20:12 crc kubenswrapper[4788]: I1010 17:20:12.212648 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 17:20:12 crc kubenswrapper[4788]: I1010 17:20:12.213362 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 10 17:20:15 crc kubenswrapper[4788]: I1010 17:20:15.225059 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 10 17:20:15 crc kubenswrapper[4788]: I1010 17:20:15.234331 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:20:15 crc kubenswrapper[4788]: E1010 17:20:15.234608 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:20:15 crc kubenswrapper[4788]: I1010 17:20:15.271932 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 10 17:20:15 crc kubenswrapper[4788]: I1010 17:20:15.929013 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 10 17:20:16 crc kubenswrapper[4788]: I1010 17:20:16.318384 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 10 17:20:16 crc kubenswrapper[4788]: I1010 17:20:16.383481 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 10 17:20:17 crc kubenswrapper[4788]: I1010 17:20:17.213122 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 17:20:17 crc kubenswrapper[4788]: I1010 17:20:17.213330 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 10 17:20:18 crc kubenswrapper[4788]: I1010 17:20:18.255384 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="19ed4b86-3570-4e57-a9f8-44e17a81beda" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.197:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 17:20:18 crc kubenswrapper[4788]: I1010 17:20:18.256226 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="19ed4b86-3570-4e57-a9f8-44e17a81beda" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.197:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 17:20:18 crc kubenswrapper[4788]: I1010 17:20:18.685914 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 17:20:18 crc kubenswrapper[4788]: I1010 17:20:18.686348 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 10 17:20:19 crc kubenswrapper[4788]: I1010 17:20:19.768500 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2029803d-4cea-46d6-af6c-7a0b17a4733c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 17:20:19 crc kubenswrapper[4788]: I1010 17:20:19.769021 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2029803d-4cea-46d6-af6c-7a0b17a4733c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 10 17:20:21 crc kubenswrapper[4788]: I1010 17:20:21.719817 4788 scope.go:117] "RemoveContainer" containerID="1cc7b83df429ca373bc48fe1d50a2e0e0360637ec494af9dda855dc6bda4e0f3" Oct 10 17:20:21 crc kubenswrapper[4788]: I1010 17:20:21.749466 4788 scope.go:117] "RemoveContainer" containerID="9f762d634176aaaa4210b783680d5baf3c7c6137f69bc900ce4efa3acb49dca4" Oct 10 17:20:27 crc kubenswrapper[4788]: I1010 17:20:27.215060 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 17:20:27 crc kubenswrapper[4788]: I1010 17:20:27.215707 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 10 17:20:27 crc kubenswrapper[4788]: I1010 17:20:27.217910 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 17:20:27 crc kubenswrapper[4788]: I1010 17:20:27.218307 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 10 17:20:27 crc kubenswrapper[4788]: I1010 17:20:27.233969 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:20:27 crc kubenswrapper[4788]: E1010 17:20:27.234341 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:20:28 crc kubenswrapper[4788]: I1010 17:20:28.691771 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 17:20:28 crc kubenswrapper[4788]: I1010 17:20:28.692566 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 10 17:20:28 crc kubenswrapper[4788]: I1010 17:20:28.694077 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 17:20:28 crc kubenswrapper[4788]: I1010 17:20:28.694127 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 10 17:20:28 crc kubenswrapper[4788]: I1010 17:20:28.698659 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 17:20:28 crc kubenswrapper[4788]: I1010 17:20:28.700945 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.896645 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk"] Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.898833 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.901374 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.901634 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.901832 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.902666 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.905648 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.908267 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.909804 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk"] Oct 10 17:20:29 crc kubenswrapper[4788]: I1010 17:20:29.912384 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-mw4pt" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.072718 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.072784 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.072904 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.072939 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.073181 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.073351 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.073406 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7xvq\" (UniqueName: \"kubernetes.io/projected/7d34aa65-68bf-476c-9b37-84a544e0eb4b-kube-api-access-s7xvq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.073446 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.073493 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.073751 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.073820 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176251 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176307 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176344 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176366 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176400 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176416 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176491 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176540 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176561 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7xvq\" (UniqueName: \"kubernetes.io/projected/7d34aa65-68bf-476c-9b37-84a544e0eb4b-kube-api-access-s7xvq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176578 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.176598 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.178189 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.178464 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.181973 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.183268 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.183402 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.183384 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.183417 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.191741 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.191909 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.192457 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.194291 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7xvq\" (UniqueName: \"kubernetes.io/projected/7d34aa65-68bf-476c-9b37-84a544e0eb4b-kube-api-access-s7xvq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.247637 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.811123 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 17:20:30 crc kubenswrapper[4788]: I1010 17:20:30.816076 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk"] Oct 10 17:20:31 crc kubenswrapper[4788]: I1010 17:20:31.065335 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" event={"ID":"7d34aa65-68bf-476c-9b37-84a544e0eb4b","Type":"ContainerStarted","Data":"3c3f1f3fb873a2f4099185391ef9ff976d5fb0b398fba68b986480634ab8e55b"} Oct 10 17:20:32 crc kubenswrapper[4788]: I1010 17:20:32.081540 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" event={"ID":"7d34aa65-68bf-476c-9b37-84a544e0eb4b","Type":"ContainerStarted","Data":"46760f25f9e24c93869319f79219a73141655b945d7b2d43ef50b3abe18b2c22"} Oct 10 17:20:32 crc kubenswrapper[4788]: I1010 17:20:32.108093 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" podStartSLOduration=2.9633180919999997 podStartE2EDuration="3.108074171s" podCreationTimestamp="2025-10-10 17:20:29 +0000 UTC" firstStartedPulling="2025-10-10 17:20:30.810906584 +0000 UTC m=+9333.260622132" lastFinishedPulling="2025-10-10 17:20:30.955662663 +0000 UTC m=+9333.405378211" observedRunningTime="2025-10-10 17:20:32.10728358 +0000 UTC m=+9334.556999148" watchObservedRunningTime="2025-10-10 17:20:32.108074171 +0000 UTC m=+9334.557789729" Oct 10 17:20:39 crc kubenswrapper[4788]: I1010 17:20:39.234316 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:20:39 crc kubenswrapper[4788]: E1010 17:20:39.235925 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.476111 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zd7kn"] Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.479615 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.505982 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zd7kn"] Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.580545 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-catalog-content\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.580602 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tq57\" (UniqueName: \"kubernetes.io/projected/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-kube-api-access-7tq57\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.580806 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-utilities\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.683229 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-utilities\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.683422 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-catalog-content\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.683450 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tq57\" (UniqueName: \"kubernetes.io/projected/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-kube-api-access-7tq57\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.684307 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-utilities\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.684332 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-catalog-content\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.708341 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tq57\" (UniqueName: \"kubernetes.io/projected/0edf0140-7a81-4ce5-bbb2-3cfebd76db36-kube-api-access-7tq57\") pod \"community-operators-zd7kn\" (UID: \"0edf0140-7a81-4ce5-bbb2-3cfebd76db36\") " pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:45 crc kubenswrapper[4788]: I1010 17:20:45.803111 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:46 crc kubenswrapper[4788]: I1010 17:20:46.307682 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zd7kn"] Oct 10 17:20:47 crc kubenswrapper[4788]: I1010 17:20:47.303408 4788 generic.go:334] "Generic (PLEG): container finished" podID="0edf0140-7a81-4ce5-bbb2-3cfebd76db36" containerID="1b91b9c2c4e86ff571d33eb1eb55fb792946c98a3f84a8dccf9006a20e18f11f" exitCode=0 Oct 10 17:20:47 crc kubenswrapper[4788]: I1010 17:20:47.303525 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zd7kn" event={"ID":"0edf0140-7a81-4ce5-bbb2-3cfebd76db36","Type":"ContainerDied","Data":"1b91b9c2c4e86ff571d33eb1eb55fb792946c98a3f84a8dccf9006a20e18f11f"} Oct 10 17:20:47 crc kubenswrapper[4788]: I1010 17:20:47.304494 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zd7kn" event={"ID":"0edf0140-7a81-4ce5-bbb2-3cfebd76db36","Type":"ContainerStarted","Data":"7e5ac3bd4db4c4eaa0b9c5ccb62d786c60d7ec555daa9dc54d2cb431ec73fc03"} Oct 10 17:20:52 crc kubenswrapper[4788]: I1010 17:20:52.234865 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:20:52 crc kubenswrapper[4788]: E1010 17:20:52.235980 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:20:52 crc kubenswrapper[4788]: I1010 17:20:52.360212 4788 generic.go:334] "Generic (PLEG): container finished" podID="0edf0140-7a81-4ce5-bbb2-3cfebd76db36" containerID="0feb30bfca520324910c77503bc7cc90c23b9f95a9598db8d9f2845fd3362d1c" exitCode=0 Oct 10 17:20:52 crc kubenswrapper[4788]: I1010 17:20:52.360351 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zd7kn" event={"ID":"0edf0140-7a81-4ce5-bbb2-3cfebd76db36","Type":"ContainerDied","Data":"0feb30bfca520324910c77503bc7cc90c23b9f95a9598db8d9f2845fd3362d1c"} Oct 10 17:20:53 crc kubenswrapper[4788]: I1010 17:20:53.374648 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zd7kn" event={"ID":"0edf0140-7a81-4ce5-bbb2-3cfebd76db36","Type":"ContainerStarted","Data":"69601085279ff39b59c2e0f0c7b645a6680fcb89691d0e4a7502ec643d4b1455"} Oct 10 17:20:53 crc kubenswrapper[4788]: I1010 17:20:53.404252 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zd7kn" podStartSLOduration=2.951034647 podStartE2EDuration="8.404229932s" podCreationTimestamp="2025-10-10 17:20:45 +0000 UTC" firstStartedPulling="2025-10-10 17:20:47.306451828 +0000 UTC m=+9349.756167396" lastFinishedPulling="2025-10-10 17:20:52.759647133 +0000 UTC m=+9355.209362681" observedRunningTime="2025-10-10 17:20:53.396315863 +0000 UTC m=+9355.846031431" watchObservedRunningTime="2025-10-10 17:20:53.404229932 +0000 UTC m=+9355.853945500" Oct 10 17:20:55 crc kubenswrapper[4788]: I1010 17:20:55.804450 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:55 crc kubenswrapper[4788]: I1010 17:20:55.804893 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:20:55 crc kubenswrapper[4788]: I1010 17:20:55.881657 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:21:05 crc kubenswrapper[4788]: I1010 17:21:05.870021 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zd7kn" Oct 10 17:21:05 crc kubenswrapper[4788]: I1010 17:21:05.959275 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zd7kn"] Oct 10 17:21:06 crc kubenswrapper[4788]: I1010 17:21:06.027471 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-clmxg"] Oct 10 17:21:06 crc kubenswrapper[4788]: I1010 17:21:06.027764 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-clmxg" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerName="registry-server" containerID="cri-o://bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608" gracePeriod=2 Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.063091 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clmxg" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.229460 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-utilities\") pod \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.229740 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t7nt\" (UniqueName: \"kubernetes.io/projected/8239a070-d3f5-4059-ac54-2b6c5b8d899e-kube-api-access-6t7nt\") pod \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.229805 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-catalog-content\") pod \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\" (UID: \"8239a070-d3f5-4059-ac54-2b6c5b8d899e\") " Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.230728 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-utilities" (OuterVolumeSpecName: "utilities") pod "8239a070-d3f5-4059-ac54-2b6c5b8d899e" (UID: "8239a070-d3f5-4059-ac54-2b6c5b8d899e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.235435 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.237420 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8239a070-d3f5-4059-ac54-2b6c5b8d899e-kube-api-access-6t7nt" (OuterVolumeSpecName: "kube-api-access-6t7nt") pod "8239a070-d3f5-4059-ac54-2b6c5b8d899e" (UID: "8239a070-d3f5-4059-ac54-2b6c5b8d899e"). InnerVolumeSpecName "kube-api-access-6t7nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.298038 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8239a070-d3f5-4059-ac54-2b6c5b8d899e" (UID: "8239a070-d3f5-4059-ac54-2b6c5b8d899e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.333638 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t7nt\" (UniqueName: \"kubernetes.io/projected/8239a070-d3f5-4059-ac54-2b6c5b8d899e-kube-api-access-6t7nt\") on node \"crc\" DevicePath \"\"" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.334243 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.334402 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8239a070-d3f5-4059-ac54-2b6c5b8d899e-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.545726 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"bba6c822e16f9f3164fb7d9f4efbff28b96355f6575bdb660154498429edf004"} Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.548440 4788 generic.go:334] "Generic (PLEG): container finished" podID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerID="bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608" exitCode=0 Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.548504 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clmxg" event={"ID":"8239a070-d3f5-4059-ac54-2b6c5b8d899e","Type":"ContainerDied","Data":"bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608"} Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.548534 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clmxg" event={"ID":"8239a070-d3f5-4059-ac54-2b6c5b8d899e","Type":"ContainerDied","Data":"64e43ff852ddf9345a3437bd689079f2bf5dcaba6da2ba182c129d1aa0973386"} Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.548554 4788 scope.go:117] "RemoveContainer" containerID="bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.548750 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clmxg" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.576033 4788 scope.go:117] "RemoveContainer" containerID="5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.609725 4788 scope.go:117] "RemoveContainer" containerID="412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.612776 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-clmxg"] Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.628054 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-clmxg"] Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.683684 4788 scope.go:117] "RemoveContainer" containerID="bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608" Oct 10 17:21:07 crc kubenswrapper[4788]: E1010 17:21:07.684199 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608\": container with ID starting with bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608 not found: ID does not exist" containerID="bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.684341 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608"} err="failed to get container status \"bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608\": rpc error: code = NotFound desc = could not find container \"bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608\": container with ID starting with bb2dc21814027d7345c586c8b6603d6b621cc3e5f8ed7307ac4e1f2e243a0608 not found: ID does not exist" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.684451 4788 scope.go:117] "RemoveContainer" containerID="5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909" Oct 10 17:21:07 crc kubenswrapper[4788]: E1010 17:21:07.684879 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909\": container with ID starting with 5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909 not found: ID does not exist" containerID="5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.684919 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909"} err="failed to get container status \"5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909\": rpc error: code = NotFound desc = could not find container \"5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909\": container with ID starting with 5be490c9d4f455af72953071a04ee00041a00d7733d831ff923b6fef21ce2909 not found: ID does not exist" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.684983 4788 scope.go:117] "RemoveContainer" containerID="412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48" Oct 10 17:21:07 crc kubenswrapper[4788]: E1010 17:21:07.685274 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48\": container with ID starting with 412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48 not found: ID does not exist" containerID="412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48" Oct 10 17:21:07 crc kubenswrapper[4788]: I1010 17:21:07.685363 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48"} err="failed to get container status \"412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48\": rpc error: code = NotFound desc = could not find container \"412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48\": container with ID starting with 412aa42375d896aa59c98cf130a8d8b1704de6a5d9b6b7ec6dfe1003569b0d48 not found: ID does not exist" Oct 10 17:21:08 crc kubenswrapper[4788]: I1010 17:21:08.282233 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" path="/var/lib/kubelet/pods/8239a070-d3f5-4059-ac54-2b6c5b8d899e/volumes" Oct 10 17:23:29 crc kubenswrapper[4788]: I1010 17:23:29.406608 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:23:29 crc kubenswrapper[4788]: I1010 17:23:29.407347 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:23:59 crc kubenswrapper[4788]: I1010 17:23:59.405801 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:23:59 crc kubenswrapper[4788]: I1010 17:23:59.406283 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.003621 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwq9"] Oct 10 17:24:24 crc kubenswrapper[4788]: E1010 17:24:24.004709 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerName="extract-utilities" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.004723 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerName="extract-utilities" Oct 10 17:24:24 crc kubenswrapper[4788]: E1010 17:24:24.004741 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerName="registry-server" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.004747 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerName="registry-server" Oct 10 17:24:24 crc kubenswrapper[4788]: E1010 17:24:24.004780 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerName="extract-content" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.004786 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerName="extract-content" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.005001 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="8239a070-d3f5-4059-ac54-2b6c5b8d899e" containerName="registry-server" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.006784 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.018405 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwq9"] Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.102253 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktn87\" (UniqueName: \"kubernetes.io/projected/c17560f5-d905-4ad5-b047-47be5e369f19-kube-api-access-ktn87\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.102379 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-catalog-content\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.102418 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-utilities\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.204898 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktn87\" (UniqueName: \"kubernetes.io/projected/c17560f5-d905-4ad5-b047-47be5e369f19-kube-api-access-ktn87\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.204992 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-catalog-content\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.205018 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-utilities\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.205683 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-utilities\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.205751 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-catalog-content\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.228403 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktn87\" (UniqueName: \"kubernetes.io/projected/c17560f5-d905-4ad5-b047-47be5e369f19-kube-api-access-ktn87\") pod \"redhat-marketplace-2dwq9\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.341374 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:24 crc kubenswrapper[4788]: I1010 17:24:24.868110 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwq9"] Oct 10 17:24:25 crc kubenswrapper[4788]: I1010 17:24:25.826775 4788 generic.go:334] "Generic (PLEG): container finished" podID="c17560f5-d905-4ad5-b047-47be5e369f19" containerID="0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239" exitCode=0 Oct 10 17:24:25 crc kubenswrapper[4788]: I1010 17:24:25.826880 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwq9" event={"ID":"c17560f5-d905-4ad5-b047-47be5e369f19","Type":"ContainerDied","Data":"0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239"} Oct 10 17:24:25 crc kubenswrapper[4788]: I1010 17:24:25.827189 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwq9" event={"ID":"c17560f5-d905-4ad5-b047-47be5e369f19","Type":"ContainerStarted","Data":"d4d72c5047510ff959f83fefb1399bdda5341e7a171d35e3e0497e1217ee1b6e"} Oct 10 17:24:27 crc kubenswrapper[4788]: I1010 17:24:27.863594 4788 generic.go:334] "Generic (PLEG): container finished" podID="c17560f5-d905-4ad5-b047-47be5e369f19" containerID="11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf" exitCode=0 Oct 10 17:24:27 crc kubenswrapper[4788]: I1010 17:24:27.863680 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwq9" event={"ID":"c17560f5-d905-4ad5-b047-47be5e369f19","Type":"ContainerDied","Data":"11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf"} Oct 10 17:24:28 crc kubenswrapper[4788]: I1010 17:24:28.879615 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwq9" event={"ID":"c17560f5-d905-4ad5-b047-47be5e369f19","Type":"ContainerStarted","Data":"369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194"} Oct 10 17:24:28 crc kubenswrapper[4788]: I1010 17:24:28.918851 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2dwq9" podStartSLOduration=3.388506597 podStartE2EDuration="5.918819556s" podCreationTimestamp="2025-10-10 17:24:23 +0000 UTC" firstStartedPulling="2025-10-10 17:24:25.837254564 +0000 UTC m=+9568.286970112" lastFinishedPulling="2025-10-10 17:24:28.367567513 +0000 UTC m=+9570.817283071" observedRunningTime="2025-10-10 17:24:28.905700121 +0000 UTC m=+9571.355415669" watchObservedRunningTime="2025-10-10 17:24:28.918819556 +0000 UTC m=+9571.368535104" Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.406825 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.406907 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.406961 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.408006 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bba6c822e16f9f3164fb7d9f4efbff28b96355f6575bdb660154498429edf004"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.408066 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://bba6c822e16f9f3164fb7d9f4efbff28b96355f6575bdb660154498429edf004" gracePeriod=600 Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.898059 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="bba6c822e16f9f3164fb7d9f4efbff28b96355f6575bdb660154498429edf004" exitCode=0 Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.898202 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"bba6c822e16f9f3164fb7d9f4efbff28b96355f6575bdb660154498429edf004"} Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.898734 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b"} Oct 10 17:24:29 crc kubenswrapper[4788]: I1010 17:24:29.898791 4788 scope.go:117] "RemoveContainer" containerID="54991816236e07e0a3feb920eb2df6599266b35520b449ed7f3059f55baa3ca0" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.588103 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nsxpv"] Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.592514 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.605874 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nsxpv"] Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.654955 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-utilities\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.655083 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rshkh\" (UniqueName: \"kubernetes.io/projected/18f1e755-00a8-42c4-a136-657f6c804267-kube-api-access-rshkh\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.655203 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-catalog-content\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.757275 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-catalog-content\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.757382 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-utilities\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.757502 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rshkh\" (UniqueName: \"kubernetes.io/projected/18f1e755-00a8-42c4-a136-657f6c804267-kube-api-access-rshkh\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.757892 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-catalog-content\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.757960 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-utilities\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.797706 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rshkh\" (UniqueName: \"kubernetes.io/projected/18f1e755-00a8-42c4-a136-657f6c804267-kube-api-access-rshkh\") pod \"redhat-operators-nsxpv\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:32 crc kubenswrapper[4788]: I1010 17:24:32.926035 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:33 crc kubenswrapper[4788]: I1010 17:24:33.477270 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nsxpv"] Oct 10 17:24:33 crc kubenswrapper[4788]: I1010 17:24:33.963370 4788 generic.go:334] "Generic (PLEG): container finished" podID="18f1e755-00a8-42c4-a136-657f6c804267" containerID="be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376" exitCode=0 Oct 10 17:24:33 crc kubenswrapper[4788]: I1010 17:24:33.963496 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsxpv" event={"ID":"18f1e755-00a8-42c4-a136-657f6c804267","Type":"ContainerDied","Data":"be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376"} Oct 10 17:24:33 crc kubenswrapper[4788]: I1010 17:24:33.963843 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsxpv" event={"ID":"18f1e755-00a8-42c4-a136-657f6c804267","Type":"ContainerStarted","Data":"9a5bb137f082dfe0af43283fe9c11367916a779cc90dbed6e11fbb84bfdf97ce"} Oct 10 17:24:34 crc kubenswrapper[4788]: I1010 17:24:34.342362 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:34 crc kubenswrapper[4788]: I1010 17:24:34.342792 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:34 crc kubenswrapper[4788]: I1010 17:24:34.435042 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:34 crc kubenswrapper[4788]: I1010 17:24:34.976686 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsxpv" event={"ID":"18f1e755-00a8-42c4-a136-657f6c804267","Type":"ContainerStarted","Data":"b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd"} Oct 10 17:24:35 crc kubenswrapper[4788]: I1010 17:24:35.028587 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:36 crc kubenswrapper[4788]: I1010 17:24:36.762220 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwq9"] Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.056159 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2dwq9" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" containerName="registry-server" containerID="cri-o://369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194" gracePeriod=2 Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.641858 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.793924 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-utilities\") pod \"c17560f5-d905-4ad5-b047-47be5e369f19\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.793981 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-catalog-content\") pod \"c17560f5-d905-4ad5-b047-47be5e369f19\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.794192 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktn87\" (UniqueName: \"kubernetes.io/projected/c17560f5-d905-4ad5-b047-47be5e369f19-kube-api-access-ktn87\") pod \"c17560f5-d905-4ad5-b047-47be5e369f19\" (UID: \"c17560f5-d905-4ad5-b047-47be5e369f19\") " Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.794814 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-utilities" (OuterVolumeSpecName: "utilities") pod "c17560f5-d905-4ad5-b047-47be5e369f19" (UID: "c17560f5-d905-4ad5-b047-47be5e369f19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.800562 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c17560f5-d905-4ad5-b047-47be5e369f19-kube-api-access-ktn87" (OuterVolumeSpecName: "kube-api-access-ktn87") pod "c17560f5-d905-4ad5-b047-47be5e369f19" (UID: "c17560f5-d905-4ad5-b047-47be5e369f19"). InnerVolumeSpecName "kube-api-access-ktn87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.805720 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c17560f5-d905-4ad5-b047-47be5e369f19" (UID: "c17560f5-d905-4ad5-b047-47be5e369f19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.896688 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.897020 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c17560f5-d905-4ad5-b047-47be5e369f19-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:24:37 crc kubenswrapper[4788]: I1010 17:24:37.897035 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktn87\" (UniqueName: \"kubernetes.io/projected/c17560f5-d905-4ad5-b047-47be5e369f19-kube-api-access-ktn87\") on node \"crc\" DevicePath \"\"" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.071834 4788 generic.go:334] "Generic (PLEG): container finished" podID="18f1e755-00a8-42c4-a136-657f6c804267" containerID="b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd" exitCode=0 Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.071945 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsxpv" event={"ID":"18f1e755-00a8-42c4-a136-657f6c804267","Type":"ContainerDied","Data":"b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd"} Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.075629 4788 generic.go:334] "Generic (PLEG): container finished" podID="c17560f5-d905-4ad5-b047-47be5e369f19" containerID="369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194" exitCode=0 Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.075666 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwq9" event={"ID":"c17560f5-d905-4ad5-b047-47be5e369f19","Type":"ContainerDied","Data":"369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194"} Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.075690 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dwq9" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.075707 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwq9" event={"ID":"c17560f5-d905-4ad5-b047-47be5e369f19","Type":"ContainerDied","Data":"d4d72c5047510ff959f83fefb1399bdda5341e7a171d35e3e0497e1217ee1b6e"} Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.075728 4788 scope.go:117] "RemoveContainer" containerID="369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.119649 4788 scope.go:117] "RemoveContainer" containerID="11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.124703 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwq9"] Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.134131 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwq9"] Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.154890 4788 scope.go:117] "RemoveContainer" containerID="0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.216956 4788 scope.go:117] "RemoveContainer" containerID="369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194" Oct 10 17:24:38 crc kubenswrapper[4788]: E1010 17:24:38.217584 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194\": container with ID starting with 369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194 not found: ID does not exist" containerID="369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.217617 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194"} err="failed to get container status \"369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194\": rpc error: code = NotFound desc = could not find container \"369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194\": container with ID starting with 369c4bf21d21cc0120f30f816a8932716a04337d0e6e0aefae75595562389194 not found: ID does not exist" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.217640 4788 scope.go:117] "RemoveContainer" containerID="11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf" Oct 10 17:24:38 crc kubenswrapper[4788]: E1010 17:24:38.218922 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf\": container with ID starting with 11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf not found: ID does not exist" containerID="11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.218972 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf"} err="failed to get container status \"11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf\": rpc error: code = NotFound desc = could not find container \"11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf\": container with ID starting with 11f21ce6c26871755a1252c9279baedb73d4f023dbf1e7e25d609218461692bf not found: ID does not exist" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.218992 4788 scope.go:117] "RemoveContainer" containerID="0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239" Oct 10 17:24:38 crc kubenswrapper[4788]: E1010 17:24:38.220454 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239\": container with ID starting with 0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239 not found: ID does not exist" containerID="0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.220488 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239"} err="failed to get container status \"0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239\": rpc error: code = NotFound desc = could not find container \"0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239\": container with ID starting with 0d788ef9e1fba6104683849c4accf444513fc4ccf267b2e85c1c4e8fea012239 not found: ID does not exist" Oct 10 17:24:38 crc kubenswrapper[4788]: I1010 17:24:38.250111 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" path="/var/lib/kubelet/pods/c17560f5-d905-4ad5-b047-47be5e369f19/volumes" Oct 10 17:24:39 crc kubenswrapper[4788]: I1010 17:24:39.094489 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsxpv" event={"ID":"18f1e755-00a8-42c4-a136-657f6c804267","Type":"ContainerStarted","Data":"ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d"} Oct 10 17:24:39 crc kubenswrapper[4788]: I1010 17:24:39.144346 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nsxpv" podStartSLOduration=2.630102451 podStartE2EDuration="7.144317282s" podCreationTimestamp="2025-10-10 17:24:32 +0000 UTC" firstStartedPulling="2025-10-10 17:24:33.967270793 +0000 UTC m=+9576.416986341" lastFinishedPulling="2025-10-10 17:24:38.481485624 +0000 UTC m=+9580.931201172" observedRunningTime="2025-10-10 17:24:39.120723441 +0000 UTC m=+9581.570438999" watchObservedRunningTime="2025-10-10 17:24:39.144317282 +0000 UTC m=+9581.594032860" Oct 10 17:24:42 crc kubenswrapper[4788]: I1010 17:24:42.926710 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:42 crc kubenswrapper[4788]: I1010 17:24:42.928300 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:43 crc kubenswrapper[4788]: I1010 17:24:43.997859 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nsxpv" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="registry-server" probeResult="failure" output=< Oct 10 17:24:43 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 17:24:43 crc kubenswrapper[4788]: > Oct 10 17:24:52 crc kubenswrapper[4788]: I1010 17:24:52.989089 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:53 crc kubenswrapper[4788]: I1010 17:24:53.049748 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:53 crc kubenswrapper[4788]: I1010 17:24:53.250563 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nsxpv"] Oct 10 17:24:54 crc kubenswrapper[4788]: I1010 17:24:54.292484 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nsxpv" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="registry-server" containerID="cri-o://ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d" gracePeriod=2 Oct 10 17:24:54 crc kubenswrapper[4788]: I1010 17:24:54.971942 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.053245 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-catalog-content\") pod \"18f1e755-00a8-42c4-a136-657f6c804267\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.053316 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rshkh\" (UniqueName: \"kubernetes.io/projected/18f1e755-00a8-42c4-a136-657f6c804267-kube-api-access-rshkh\") pod \"18f1e755-00a8-42c4-a136-657f6c804267\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.053432 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-utilities\") pod \"18f1e755-00a8-42c4-a136-657f6c804267\" (UID: \"18f1e755-00a8-42c4-a136-657f6c804267\") " Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.054570 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-utilities" (OuterVolumeSpecName: "utilities") pod "18f1e755-00a8-42c4-a136-657f6c804267" (UID: "18f1e755-00a8-42c4-a136-657f6c804267"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.098711 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f1e755-00a8-42c4-a136-657f6c804267-kube-api-access-rshkh" (OuterVolumeSpecName: "kube-api-access-rshkh") pod "18f1e755-00a8-42c4-a136-657f6c804267" (UID: "18f1e755-00a8-42c4-a136-657f6c804267"). InnerVolumeSpecName "kube-api-access-rshkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.157278 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rshkh\" (UniqueName: \"kubernetes.io/projected/18f1e755-00a8-42c4-a136-657f6c804267-kube-api-access-rshkh\") on node \"crc\" DevicePath \"\"" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.157308 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.178533 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18f1e755-00a8-42c4-a136-657f6c804267" (UID: "18f1e755-00a8-42c4-a136-657f6c804267"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.262101 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18f1e755-00a8-42c4-a136-657f6c804267-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.310061 4788 generic.go:334] "Generic (PLEG): container finished" podID="18f1e755-00a8-42c4-a136-657f6c804267" containerID="ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d" exitCode=0 Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.310167 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsxpv" event={"ID":"18f1e755-00a8-42c4-a136-657f6c804267","Type":"ContainerDied","Data":"ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d"} Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.310356 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsxpv" event={"ID":"18f1e755-00a8-42c4-a136-657f6c804267","Type":"ContainerDied","Data":"9a5bb137f082dfe0af43283fe9c11367916a779cc90dbed6e11fbb84bfdf97ce"} Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.310392 4788 scope.go:117] "RemoveContainer" containerID="ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.310270 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsxpv" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.345550 4788 scope.go:117] "RemoveContainer" containerID="b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.351804 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nsxpv"] Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.360854 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nsxpv"] Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.383755 4788 scope.go:117] "RemoveContainer" containerID="be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.417811 4788 scope.go:117] "RemoveContainer" containerID="ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d" Oct 10 17:24:55 crc kubenswrapper[4788]: E1010 17:24:55.418349 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d\": container with ID starting with ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d not found: ID does not exist" containerID="ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.418398 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d"} err="failed to get container status \"ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d\": rpc error: code = NotFound desc = could not find container \"ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d\": container with ID starting with ba535dc47d7a6a72e7294c82578d15ef10c39856b7291663d4199aa501069e6d not found: ID does not exist" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.418429 4788 scope.go:117] "RemoveContainer" containerID="b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd" Oct 10 17:24:55 crc kubenswrapper[4788]: E1010 17:24:55.418886 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd\": container with ID starting with b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd not found: ID does not exist" containerID="b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.418943 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd"} err="failed to get container status \"b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd\": rpc error: code = NotFound desc = could not find container \"b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd\": container with ID starting with b5ad4b6f1587f1de888179a13b85fb85227d5cb917c9ddd036b025c0ac7668dd not found: ID does not exist" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.418979 4788 scope.go:117] "RemoveContainer" containerID="be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376" Oct 10 17:24:55 crc kubenswrapper[4788]: E1010 17:24:55.419372 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376\": container with ID starting with be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376 not found: ID does not exist" containerID="be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376" Oct 10 17:24:55 crc kubenswrapper[4788]: I1010 17:24:55.419414 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376"} err="failed to get container status \"be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376\": rpc error: code = NotFound desc = could not find container \"be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376\": container with ID starting with be38cd08a0becc80667e4ba6745faebf537be3d3fe2d7ab9166a63f8ded22376 not found: ID does not exist" Oct 10 17:24:56 crc kubenswrapper[4788]: I1010 17:24:56.258472 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18f1e755-00a8-42c4-a136-657f6c804267" path="/var/lib/kubelet/pods/18f1e755-00a8-42c4-a136-657f6c804267/volumes" Oct 10 17:26:29 crc kubenswrapper[4788]: I1010 17:26:29.406528 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:26:29 crc kubenswrapper[4788]: I1010 17:26:29.407304 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:26:59 crc kubenswrapper[4788]: I1010 17:26:59.406268 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:26:59 crc kubenswrapper[4788]: I1010 17:26:59.406956 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:27:29 crc kubenswrapper[4788]: I1010 17:27:29.406423 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:27:29 crc kubenswrapper[4788]: I1010 17:27:29.407118 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:27:29 crc kubenswrapper[4788]: I1010 17:27:29.407201 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 17:27:29 crc kubenswrapper[4788]: I1010 17:27:29.408159 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 17:27:29 crc kubenswrapper[4788]: I1010 17:27:29.408211 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" gracePeriod=600 Oct 10 17:27:29 crc kubenswrapper[4788]: E1010 17:27:29.540238 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:27:30 crc kubenswrapper[4788]: I1010 17:27:30.282516 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" exitCode=0 Oct 10 17:27:30 crc kubenswrapper[4788]: I1010 17:27:30.282557 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b"} Oct 10 17:27:30 crc kubenswrapper[4788]: I1010 17:27:30.283003 4788 scope.go:117] "RemoveContainer" containerID="bba6c822e16f9f3164fb7d9f4efbff28b96355f6575bdb660154498429edf004" Oct 10 17:27:30 crc kubenswrapper[4788]: I1010 17:27:30.283840 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:27:30 crc kubenswrapper[4788]: E1010 17:27:30.284230 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:27:44 crc kubenswrapper[4788]: I1010 17:27:44.235498 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:27:44 crc kubenswrapper[4788]: E1010 17:27:44.236599 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:27:55 crc kubenswrapper[4788]: I1010 17:27:55.234885 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:27:55 crc kubenswrapper[4788]: E1010 17:27:55.235498 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:28:07 crc kubenswrapper[4788]: I1010 17:28:07.234326 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:28:07 crc kubenswrapper[4788]: E1010 17:28:07.235157 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:28:20 crc kubenswrapper[4788]: I1010 17:28:20.234162 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:28:20 crc kubenswrapper[4788]: E1010 17:28:20.234735 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:28:34 crc kubenswrapper[4788]: I1010 17:28:34.235284 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:28:34 crc kubenswrapper[4788]: E1010 17:28:34.236997 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:28:47 crc kubenswrapper[4788]: I1010 17:28:47.235128 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:28:47 crc kubenswrapper[4788]: E1010 17:28:47.235891 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:28:58 crc kubenswrapper[4788]: I1010 17:28:58.243869 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:28:58 crc kubenswrapper[4788]: E1010 17:28:58.245103 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:29:11 crc kubenswrapper[4788]: I1010 17:29:11.235005 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:29:11 crc kubenswrapper[4788]: E1010 17:29:11.236126 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:29:22 crc kubenswrapper[4788]: I1010 17:29:22.234967 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:29:22 crc kubenswrapper[4788]: E1010 17:29:22.235860 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:29:34 crc kubenswrapper[4788]: I1010 17:29:34.234064 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:29:34 crc kubenswrapper[4788]: E1010 17:29:34.234968 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:29:49 crc kubenswrapper[4788]: I1010 17:29:49.234007 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:29:49 crc kubenswrapper[4788]: E1010 17:29:49.234683 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.168458 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r"] Oct 10 17:30:00 crc kubenswrapper[4788]: E1010 17:30:00.169544 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" containerName="extract-content" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.169561 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" containerName="extract-content" Oct 10 17:30:00 crc kubenswrapper[4788]: E1010 17:30:00.169581 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" containerName="registry-server" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.169587 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" containerName="registry-server" Oct 10 17:30:00 crc kubenswrapper[4788]: E1010 17:30:00.169603 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" containerName="extract-utilities" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.169609 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" containerName="extract-utilities" Oct 10 17:30:00 crc kubenswrapper[4788]: E1010 17:30:00.169625 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="extract-content" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.169631 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="extract-content" Oct 10 17:30:00 crc kubenswrapper[4788]: E1010 17:30:00.169642 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="extract-utilities" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.169648 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="extract-utilities" Oct 10 17:30:00 crc kubenswrapper[4788]: E1010 17:30:00.169673 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="registry-server" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.169679 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="registry-server" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.169893 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="c17560f5-d905-4ad5-b047-47be5e369f19" containerName="registry-server" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.169911 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="18f1e755-00a8-42c4-a136-657f6c804267" containerName="registry-server" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.172053 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.174755 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.175118 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.202209 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r"] Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.314416 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-config-volume\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.314458 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-secret-volume\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.314558 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4xs7\" (UniqueName: \"kubernetes.io/projected/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-kube-api-access-x4xs7\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.416193 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4xs7\" (UniqueName: \"kubernetes.io/projected/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-kube-api-access-x4xs7\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.416341 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-config-volume\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.416364 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-secret-volume\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.417192 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-config-volume\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.423358 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-secret-volume\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.434043 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4xs7\" (UniqueName: \"kubernetes.io/projected/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-kube-api-access-x4xs7\") pod \"collect-profiles-29335290-x6m4r\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:00 crc kubenswrapper[4788]: I1010 17:30:00.506123 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:01 crc kubenswrapper[4788]: I1010 17:30:01.004618 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r"] Oct 10 17:30:01 crc kubenswrapper[4788]: I1010 17:30:01.080600 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" event={"ID":"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b","Type":"ContainerStarted","Data":"4b91733638ed8301b1fb312b6a372ae7db24895319508afc5a527df8df8bc6de"} Oct 10 17:30:02 crc kubenswrapper[4788]: I1010 17:30:02.092936 4788 generic.go:334] "Generic (PLEG): container finished" podID="0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b" containerID="cc3502586a59ea7ed4af4dec1a2bd7d02e085940cd320d73098e7cf5185b6090" exitCode=0 Oct 10 17:30:02 crc kubenswrapper[4788]: I1010 17:30:02.093002 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" event={"ID":"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b","Type":"ContainerDied","Data":"cc3502586a59ea7ed4af4dec1a2bd7d02e085940cd320d73098e7cf5185b6090"} Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.233732 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:30:03 crc kubenswrapper[4788]: E1010 17:30:03.235292 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.554517 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.689326 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-config-volume\") pod \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.689380 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-secret-volume\") pod \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.689699 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4xs7\" (UniqueName: \"kubernetes.io/projected/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-kube-api-access-x4xs7\") pod \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\" (UID: \"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b\") " Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.689997 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-config-volume" (OuterVolumeSpecName: "config-volume") pod "0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b" (UID: "0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.690446 4788 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.696014 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b" (UID: "0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.696105 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-kube-api-access-x4xs7" (OuterVolumeSpecName: "kube-api-access-x4xs7") pod "0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b" (UID: "0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b"). InnerVolumeSpecName "kube-api-access-x4xs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.796428 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4xs7\" (UniqueName: \"kubernetes.io/projected/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-kube-api-access-x4xs7\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:03 crc kubenswrapper[4788]: I1010 17:30:03.796463 4788 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:04 crc kubenswrapper[4788]: I1010 17:30:04.115448 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" event={"ID":"0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b","Type":"ContainerDied","Data":"4b91733638ed8301b1fb312b6a372ae7db24895319508afc5a527df8df8bc6de"} Oct 10 17:30:04 crc kubenswrapper[4788]: I1010 17:30:04.115519 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b91733638ed8301b1fb312b6a372ae7db24895319508afc5a527df8df8bc6de" Oct 10 17:30:04 crc kubenswrapper[4788]: I1010 17:30:04.115523 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335290-x6m4r" Oct 10 17:30:04 crc kubenswrapper[4788]: I1010 17:30:04.637422 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4"] Oct 10 17:30:04 crc kubenswrapper[4788]: I1010 17:30:04.646959 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335245-p25n4"] Oct 10 17:30:06 crc kubenswrapper[4788]: I1010 17:30:06.246506 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="523d2a02-011c-4feb-a2e5-8468d7f1e3eb" path="/var/lib/kubelet/pods/523d2a02-011c-4feb-a2e5-8468d7f1e3eb/volumes" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.115438 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ht6f7"] Oct 10 17:30:14 crc kubenswrapper[4788]: E1010 17:30:14.116542 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b" containerName="collect-profiles" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.116558 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b" containerName="collect-profiles" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.116782 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e916ad4-c83c-42f4-83bd-ac1ccdb46d5b" containerName="collect-profiles" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.118469 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.132786 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ht6f7"] Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.234419 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:30:14 crc kubenswrapper[4788]: E1010 17:30:14.235209 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.283323 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-utilities\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.283562 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-catalog-content\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.283685 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dhmf\" (UniqueName: \"kubernetes.io/projected/0476ef7e-7db8-4944-b560-79d0a516c845-kube-api-access-5dhmf\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.384862 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dhmf\" (UniqueName: \"kubernetes.io/projected/0476ef7e-7db8-4944-b560-79d0a516c845-kube-api-access-5dhmf\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.384969 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-utilities\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.385066 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-catalog-content\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.385600 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-catalog-content\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.385599 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-utilities\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.409033 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dhmf\" (UniqueName: \"kubernetes.io/projected/0476ef7e-7db8-4944-b560-79d0a516c845-kube-api-access-5dhmf\") pod \"certified-operators-ht6f7\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.448071 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:14 crc kubenswrapper[4788]: I1010 17:30:14.982388 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ht6f7"] Oct 10 17:30:15 crc kubenswrapper[4788]: I1010 17:30:15.274408 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht6f7" event={"ID":"0476ef7e-7db8-4944-b560-79d0a516c845","Type":"ContainerDied","Data":"b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb"} Oct 10 17:30:15 crc kubenswrapper[4788]: I1010 17:30:15.274129 4788 generic.go:334] "Generic (PLEG): container finished" podID="0476ef7e-7db8-4944-b560-79d0a516c845" containerID="b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb" exitCode=0 Oct 10 17:30:15 crc kubenswrapper[4788]: I1010 17:30:15.275936 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht6f7" event={"ID":"0476ef7e-7db8-4944-b560-79d0a516c845","Type":"ContainerStarted","Data":"33ef923f6edf929a195d357e7da52c44ad46b17b28f4fb7c62cd54403344acc5"} Oct 10 17:30:15 crc kubenswrapper[4788]: I1010 17:30:15.279818 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 17:30:16 crc kubenswrapper[4788]: I1010 17:30:16.292688 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht6f7" event={"ID":"0476ef7e-7db8-4944-b560-79d0a516c845","Type":"ContainerStarted","Data":"2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0"} Oct 10 17:30:17 crc kubenswrapper[4788]: I1010 17:30:17.307991 4788 generic.go:334] "Generic (PLEG): container finished" podID="0476ef7e-7db8-4944-b560-79d0a516c845" containerID="2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0" exitCode=0 Oct 10 17:30:17 crc kubenswrapper[4788]: I1010 17:30:17.308266 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht6f7" event={"ID":"0476ef7e-7db8-4944-b560-79d0a516c845","Type":"ContainerDied","Data":"2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0"} Oct 10 17:30:18 crc kubenswrapper[4788]: I1010 17:30:18.333883 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht6f7" event={"ID":"0476ef7e-7db8-4944-b560-79d0a516c845","Type":"ContainerStarted","Data":"6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a"} Oct 10 17:30:18 crc kubenswrapper[4788]: I1010 17:30:18.359426 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ht6f7" podStartSLOduration=1.8772782719999999 podStartE2EDuration="4.359406323s" podCreationTimestamp="2025-10-10 17:30:14 +0000 UTC" firstStartedPulling="2025-10-10 17:30:15.279550506 +0000 UTC m=+9917.729266064" lastFinishedPulling="2025-10-10 17:30:17.761678557 +0000 UTC m=+9920.211394115" observedRunningTime="2025-10-10 17:30:18.352357217 +0000 UTC m=+9920.802072765" watchObservedRunningTime="2025-10-10 17:30:18.359406323 +0000 UTC m=+9920.809121871" Oct 10 17:30:22 crc kubenswrapper[4788]: I1010 17:30:22.549120 4788 scope.go:117] "RemoveContainer" containerID="92a5753885667dbaaa8ca6ebaf4aada52c46271994bfe911f64568ce6a6c08f3" Oct 10 17:30:24 crc kubenswrapper[4788]: I1010 17:30:24.449586 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:24 crc kubenswrapper[4788]: I1010 17:30:24.450159 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:24 crc kubenswrapper[4788]: I1010 17:30:24.525732 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:25 crc kubenswrapper[4788]: I1010 17:30:25.489019 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:25 crc kubenswrapper[4788]: I1010 17:30:25.536895 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ht6f7"] Oct 10 17:30:27 crc kubenswrapper[4788]: I1010 17:30:27.452413 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ht6f7" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" containerName="registry-server" containerID="cri-o://6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a" gracePeriod=2 Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.046846 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.224408 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dhmf\" (UniqueName: \"kubernetes.io/projected/0476ef7e-7db8-4944-b560-79d0a516c845-kube-api-access-5dhmf\") pod \"0476ef7e-7db8-4944-b560-79d0a516c845\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.224780 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-utilities\") pod \"0476ef7e-7db8-4944-b560-79d0a516c845\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.224850 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-catalog-content\") pod \"0476ef7e-7db8-4944-b560-79d0a516c845\" (UID: \"0476ef7e-7db8-4944-b560-79d0a516c845\") " Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.227098 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-utilities" (OuterVolumeSpecName: "utilities") pod "0476ef7e-7db8-4944-b560-79d0a516c845" (UID: "0476ef7e-7db8-4944-b560-79d0a516c845"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.233842 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0476ef7e-7db8-4944-b560-79d0a516c845-kube-api-access-5dhmf" (OuterVolumeSpecName: "kube-api-access-5dhmf") pod "0476ef7e-7db8-4944-b560-79d0a516c845" (UID: "0476ef7e-7db8-4944-b560-79d0a516c845"). InnerVolumeSpecName "kube-api-access-5dhmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.243162 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:30:28 crc kubenswrapper[4788]: E1010 17:30:28.243721 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.291016 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0476ef7e-7db8-4944-b560-79d0a516c845" (UID: "0476ef7e-7db8-4944-b560-79d0a516c845"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.327363 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dhmf\" (UniqueName: \"kubernetes.io/projected/0476ef7e-7db8-4944-b560-79d0a516c845-kube-api-access-5dhmf\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.327402 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.327413 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0476ef7e-7db8-4944-b560-79d0a516c845-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.462796 4788 generic.go:334] "Generic (PLEG): container finished" podID="7d34aa65-68bf-476c-9b37-84a544e0eb4b" containerID="46760f25f9e24c93869319f79219a73141655b945d7b2d43ef50b3abe18b2c22" exitCode=0 Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.462821 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" event={"ID":"7d34aa65-68bf-476c-9b37-84a544e0eb4b","Type":"ContainerDied","Data":"46760f25f9e24c93869319f79219a73141655b945d7b2d43ef50b3abe18b2c22"} Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.464665 4788 generic.go:334] "Generic (PLEG): container finished" podID="0476ef7e-7db8-4944-b560-79d0a516c845" containerID="6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a" exitCode=0 Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.464696 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht6f7" event={"ID":"0476ef7e-7db8-4944-b560-79d0a516c845","Type":"ContainerDied","Data":"6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a"} Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.464712 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ht6f7" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.464716 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht6f7" event={"ID":"0476ef7e-7db8-4944-b560-79d0a516c845","Type":"ContainerDied","Data":"33ef923f6edf929a195d357e7da52c44ad46b17b28f4fb7c62cd54403344acc5"} Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.464739 4788 scope.go:117] "RemoveContainer" containerID="6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.518468 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ht6f7"] Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.522255 4788 scope.go:117] "RemoveContainer" containerID="2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.526752 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ht6f7"] Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.549089 4788 scope.go:117] "RemoveContainer" containerID="b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.595334 4788 scope.go:117] "RemoveContainer" containerID="6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a" Oct 10 17:30:28 crc kubenswrapper[4788]: E1010 17:30:28.595805 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a\": container with ID starting with 6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a not found: ID does not exist" containerID="6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.595835 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a"} err="failed to get container status \"6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a\": rpc error: code = NotFound desc = could not find container \"6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a\": container with ID starting with 6daf9a92c28666b83ffd730e7726baf3c9f0605e3d1daf0947e36434eb26563a not found: ID does not exist" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.595853 4788 scope.go:117] "RemoveContainer" containerID="2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0" Oct 10 17:30:28 crc kubenswrapper[4788]: E1010 17:30:28.596656 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0\": container with ID starting with 2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0 not found: ID does not exist" containerID="2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.596682 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0"} err="failed to get container status \"2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0\": rpc error: code = NotFound desc = could not find container \"2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0\": container with ID starting with 2668c8661e8495adfe81a8f5e9cd1bee1135652097d55bc18556e6e468e4a4f0 not found: ID does not exist" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.596696 4788 scope.go:117] "RemoveContainer" containerID="b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb" Oct 10 17:30:28 crc kubenswrapper[4788]: E1010 17:30:28.596997 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb\": container with ID starting with b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb not found: ID does not exist" containerID="b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb" Oct 10 17:30:28 crc kubenswrapper[4788]: I1010 17:30:28.597018 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb"} err="failed to get container status \"b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb\": rpc error: code = NotFound desc = could not find container \"b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb\": container with ID starting with b2c5843352c618ae767bfbc181f1854b83f02883ba4d4e5b3618986f58df78fb not found: ID does not exist" Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.956931 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.977740 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-0\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.977862 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-0\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.977906 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.977966 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ceph\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.978050 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-inventory\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.978085 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-1\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.978241 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-1\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.978278 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-0\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.978313 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-1\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.978445 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-combined-ca-bundle\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.978499 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7xvq\" (UniqueName: \"kubernetes.io/projected/7d34aa65-68bf-476c-9b37-84a544e0eb4b-kube-api-access-s7xvq\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.988347 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d34aa65-68bf-476c-9b37-84a544e0eb4b-kube-api-access-s7xvq" (OuterVolumeSpecName: "kube-api-access-s7xvq") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "kube-api-access-s7xvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:30:29 crc kubenswrapper[4788]: I1010 17:30:29.990057 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.000526 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ceph" (OuterVolumeSpecName: "ceph") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.036834 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.041107 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.048742 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.057503 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.063844 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.064965 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-inventory" (OuterVolumeSpecName: "inventory") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: E1010 17:30:30.068277 4788 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key podName:7d34aa65-68bf-476c-9b37-84a544e0eb4b nodeName:}" failed. No retries permitted until 2025-10-10 17:30:30.568255094 +0000 UTC m=+9933.017970642 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b") : error deleting /var/lib/kubelet/pods/7d34aa65-68bf-476c-9b37-84a544e0eb4b/volume-subpaths: remove /var/lib/kubelet/pods/7d34aa65-68bf-476c-9b37-84a544e0eb4b/volume-subpaths: no such file or directory Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.068766 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080336 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080362 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7xvq\" (UniqueName: \"kubernetes.io/projected/7d34aa65-68bf-476c-9b37-84a544e0eb4b-kube-api-access-s7xvq\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080371 4788 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080380 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080389 4788 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ceph\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080396 4788 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-inventory\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080404 4788 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080413 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080421 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.080428 4788 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7d34aa65-68bf-476c-9b37-84a544e0eb4b-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.250811 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" path="/var/lib/kubelet/pods/0476ef7e-7db8-4944-b560-79d0a516c845/volumes" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.494385 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" event={"ID":"7d34aa65-68bf-476c-9b37-84a544e0eb4b","Type":"ContainerDied","Data":"3c3f1f3fb873a2f4099185391ef9ff976d5fb0b398fba68b986480634ab8e55b"} Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.494428 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c3f1f3fb873a2f4099185391ef9ff976d5fb0b398fba68b986480634ab8e55b" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.494451 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk" Oct 10 17:30:30 crc kubenswrapper[4788]: I1010 17:30:30.589267 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key\") pod \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\" (UID: \"7d34aa65-68bf-476c-9b37-84a544e0eb4b\") " Oct 10 17:30:31 crc kubenswrapper[4788]: I1010 17:30:31.206120 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7d34aa65-68bf-476c-9b37-84a544e0eb4b" (UID: "7d34aa65-68bf-476c-9b37-84a544e0eb4b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 10 17:30:31 crc kubenswrapper[4788]: I1010 17:30:31.306311 4788 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d34aa65-68bf-476c-9b37-84a544e0eb4b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 10 17:30:41 crc kubenswrapper[4788]: I1010 17:30:41.234083 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:30:41 crc kubenswrapper[4788]: E1010 17:30:41.235084 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:30:56 crc kubenswrapper[4788]: I1010 17:30:56.235632 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:30:56 crc kubenswrapper[4788]: E1010 17:30:56.236681 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:31:10 crc kubenswrapper[4788]: I1010 17:31:10.235100 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:31:10 crc kubenswrapper[4788]: E1010 17:31:10.236076 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:31:25 crc kubenswrapper[4788]: I1010 17:31:25.233985 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:31:25 crc kubenswrapper[4788]: E1010 17:31:25.234636 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:31:38 crc kubenswrapper[4788]: I1010 17:31:38.240456 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:31:38 crc kubenswrapper[4788]: E1010 17:31:38.241094 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:31:49 crc kubenswrapper[4788]: I1010 17:31:49.234568 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:31:49 crc kubenswrapper[4788]: E1010 17:31:49.235352 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:32:01 crc kubenswrapper[4788]: I1010 17:32:01.234435 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:32:01 crc kubenswrapper[4788]: E1010 17:32:01.235341 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:32:13 crc kubenswrapper[4788]: I1010 17:32:13.235220 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:32:13 crc kubenswrapper[4788]: E1010 17:32:13.238089 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:32:28 crc kubenswrapper[4788]: I1010 17:32:28.241736 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:32:28 crc kubenswrapper[4788]: E1010 17:32:28.242472 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:32:42 crc kubenswrapper[4788]: I1010 17:32:42.236774 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:32:43 crc kubenswrapper[4788]: I1010 17:32:43.284217 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"93046d31ffbebfcb70b32194820289c869f35af289dc70cd1e47ca5fb7e4b2bf"} Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.042380 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7xjww/must-gather-wfqvz"] Oct 10 17:32:52 crc kubenswrapper[4788]: E1010 17:32:52.043962 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" containerName="registry-server" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.043981 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" containerName="registry-server" Oct 10 17:32:52 crc kubenswrapper[4788]: E1010 17:32:52.044007 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" containerName="extract-utilities" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.044015 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" containerName="extract-utilities" Oct 10 17:32:52 crc kubenswrapper[4788]: E1010 17:32:52.044030 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d34aa65-68bf-476c-9b37-84a544e0eb4b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.044041 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d34aa65-68bf-476c-9b37-84a544e0eb4b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 10 17:32:52 crc kubenswrapper[4788]: E1010 17:32:52.044076 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" containerName="extract-content" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.044085 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" containerName="extract-content" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.044396 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="0476ef7e-7db8-4944-b560-79d0a516c845" containerName="registry-server" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.044422 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d34aa65-68bf-476c-9b37-84a544e0eb4b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.046071 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.049690 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7xjww"/"kube-root-ca.crt" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.050097 4788 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7xjww"/"openshift-service-ca.crt" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.050259 4788 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7xjww"/"default-dockercfg-fp5bf" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.059524 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7xjww/must-gather-wfqvz"] Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.127325 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4nwf\" (UniqueName: \"kubernetes.io/projected/5c789567-43c2-4f52-b62d-cd6b956a4f47-kube-api-access-l4nwf\") pod \"must-gather-wfqvz\" (UID: \"5c789567-43c2-4f52-b62d-cd6b956a4f47\") " pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.127400 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5c789567-43c2-4f52-b62d-cd6b956a4f47-must-gather-output\") pod \"must-gather-wfqvz\" (UID: \"5c789567-43c2-4f52-b62d-cd6b956a4f47\") " pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.230096 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4nwf\" (UniqueName: \"kubernetes.io/projected/5c789567-43c2-4f52-b62d-cd6b956a4f47-kube-api-access-l4nwf\") pod \"must-gather-wfqvz\" (UID: \"5c789567-43c2-4f52-b62d-cd6b956a4f47\") " pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.230173 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5c789567-43c2-4f52-b62d-cd6b956a4f47-must-gather-output\") pod \"must-gather-wfqvz\" (UID: \"5c789567-43c2-4f52-b62d-cd6b956a4f47\") " pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.230739 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5c789567-43c2-4f52-b62d-cd6b956a4f47-must-gather-output\") pod \"must-gather-wfqvz\" (UID: \"5c789567-43c2-4f52-b62d-cd6b956a4f47\") " pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.803526 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4nwf\" (UniqueName: \"kubernetes.io/projected/5c789567-43c2-4f52-b62d-cd6b956a4f47-kube-api-access-l4nwf\") pod \"must-gather-wfqvz\" (UID: \"5c789567-43c2-4f52-b62d-cd6b956a4f47\") " pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:32:52 crc kubenswrapper[4788]: I1010 17:32:52.968268 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:32:53 crc kubenswrapper[4788]: I1010 17:32:53.487163 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7xjww/must-gather-wfqvz"] Oct 10 17:32:54 crc kubenswrapper[4788]: I1010 17:32:54.432263 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/must-gather-wfqvz" event={"ID":"5c789567-43c2-4f52-b62d-cd6b956a4f47","Type":"ContainerStarted","Data":"9c9deb13b3737f9561906d3c1f973c5c3c6655a47652f7a3d8690a23eefd6d38"} Oct 10 17:33:02 crc kubenswrapper[4788]: I1010 17:33:02.521969 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/must-gather-wfqvz" event={"ID":"5c789567-43c2-4f52-b62d-cd6b956a4f47","Type":"ContainerStarted","Data":"5f56aa7cdfe98aed70e4952da11818dcc32258966b2a3f9dbefd4e2dcad912fb"} Oct 10 17:33:02 crc kubenswrapper[4788]: I1010 17:33:02.522471 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/must-gather-wfqvz" event={"ID":"5c789567-43c2-4f52-b62d-cd6b956a4f47","Type":"ContainerStarted","Data":"1c3f4cf0fd24e46294ad698717488da6c2e6d637e8fe2db32e108f640e723e52"} Oct 10 17:33:02 crc kubenswrapper[4788]: I1010 17:33:02.554886 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7xjww/must-gather-wfqvz" podStartSLOduration=2.505330426 podStartE2EDuration="10.554862907s" podCreationTimestamp="2025-10-10 17:32:52 +0000 UTC" firstStartedPulling="2025-10-10 17:32:53.498308613 +0000 UTC m=+10075.948024161" lastFinishedPulling="2025-10-10 17:33:01.547841094 +0000 UTC m=+10083.997556642" observedRunningTime="2025-10-10 17:33:02.541122915 +0000 UTC m=+10084.990838473" watchObservedRunningTime="2025-10-10 17:33:02.554862907 +0000 UTC m=+10085.004578475" Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.421319 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7xjww/crc-debug-5b68h"] Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.424473 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.517598 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58f8ca7c-88c0-45ed-b957-3e5000625007-host\") pod \"crc-debug-5b68h\" (UID: \"58f8ca7c-88c0-45ed-b957-3e5000625007\") " pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.517870 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d96dm\" (UniqueName: \"kubernetes.io/projected/58f8ca7c-88c0-45ed-b957-3e5000625007-kube-api-access-d96dm\") pod \"crc-debug-5b68h\" (UID: \"58f8ca7c-88c0-45ed-b957-3e5000625007\") " pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.619772 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58f8ca7c-88c0-45ed-b957-3e5000625007-host\") pod \"crc-debug-5b68h\" (UID: \"58f8ca7c-88c0-45ed-b957-3e5000625007\") " pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.619881 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58f8ca7c-88c0-45ed-b957-3e5000625007-host\") pod \"crc-debug-5b68h\" (UID: \"58f8ca7c-88c0-45ed-b957-3e5000625007\") " pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.619985 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d96dm\" (UniqueName: \"kubernetes.io/projected/58f8ca7c-88c0-45ed-b957-3e5000625007-kube-api-access-d96dm\") pod \"crc-debug-5b68h\" (UID: \"58f8ca7c-88c0-45ed-b957-3e5000625007\") " pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.658824 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d96dm\" (UniqueName: \"kubernetes.io/projected/58f8ca7c-88c0-45ed-b957-3e5000625007-kube-api-access-d96dm\") pod \"crc-debug-5b68h\" (UID: \"58f8ca7c-88c0-45ed-b957-3e5000625007\") " pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:06 crc kubenswrapper[4788]: I1010 17:33:06.747577 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:06 crc kubenswrapper[4788]: W1010 17:33:06.858292 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58f8ca7c_88c0_45ed_b957_3e5000625007.slice/crio-53544ba9776f31487912a21c8121031d2cdf6b591498db55c511e03befba8381 WatchSource:0}: Error finding container 53544ba9776f31487912a21c8121031d2cdf6b591498db55c511e03befba8381: Status 404 returned error can't find the container with id 53544ba9776f31487912a21c8121031d2cdf6b591498db55c511e03befba8381 Oct 10 17:33:07 crc kubenswrapper[4788]: I1010 17:33:07.575373 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/crc-debug-5b68h" event={"ID":"58f8ca7c-88c0-45ed-b957-3e5000625007","Type":"ContainerStarted","Data":"53544ba9776f31487912a21c8121031d2cdf6b591498db55c511e03befba8381"} Oct 10 17:33:19 crc kubenswrapper[4788]: I1010 17:33:19.710254 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/crc-debug-5b68h" event={"ID":"58f8ca7c-88c0-45ed-b957-3e5000625007","Type":"ContainerStarted","Data":"b2b06f6839d441a349f3551457279ed2f62ed62442cb06f2b62336e513948757"} Oct 10 17:33:35 crc kubenswrapper[4788]: I1010 17:33:35.892346 4788 generic.go:334] "Generic (PLEG): container finished" podID="58f8ca7c-88c0-45ed-b957-3e5000625007" containerID="b2b06f6839d441a349f3551457279ed2f62ed62442cb06f2b62336e513948757" exitCode=0 Oct 10 17:33:35 crc kubenswrapper[4788]: I1010 17:33:35.892482 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/crc-debug-5b68h" event={"ID":"58f8ca7c-88c0-45ed-b957-3e5000625007","Type":"ContainerDied","Data":"b2b06f6839d441a349f3551457279ed2f62ed62442cb06f2b62336e513948757"} Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.040429 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.081612 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7xjww/crc-debug-5b68h"] Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.085333 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d96dm\" (UniqueName: \"kubernetes.io/projected/58f8ca7c-88c0-45ed-b957-3e5000625007-kube-api-access-d96dm\") pod \"58f8ca7c-88c0-45ed-b957-3e5000625007\" (UID: \"58f8ca7c-88c0-45ed-b957-3e5000625007\") " Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.085416 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58f8ca7c-88c0-45ed-b957-3e5000625007-host\") pod \"58f8ca7c-88c0-45ed-b957-3e5000625007\" (UID: \"58f8ca7c-88c0-45ed-b957-3e5000625007\") " Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.085499 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58f8ca7c-88c0-45ed-b957-3e5000625007-host" (OuterVolumeSpecName: "host") pod "58f8ca7c-88c0-45ed-b957-3e5000625007" (UID: "58f8ca7c-88c0-45ed-b957-3e5000625007"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.085862 4788 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58f8ca7c-88c0-45ed-b957-3e5000625007-host\") on node \"crc\" DevicePath \"\"" Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.090027 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7xjww/crc-debug-5b68h"] Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.105097 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58f8ca7c-88c0-45ed-b957-3e5000625007-kube-api-access-d96dm" (OuterVolumeSpecName: "kube-api-access-d96dm") pod "58f8ca7c-88c0-45ed-b957-3e5000625007" (UID: "58f8ca7c-88c0-45ed-b957-3e5000625007"). InnerVolumeSpecName "kube-api-access-d96dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.188669 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d96dm\" (UniqueName: \"kubernetes.io/projected/58f8ca7c-88c0-45ed-b957-3e5000625007-kube-api-access-d96dm\") on node \"crc\" DevicePath \"\"" Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.919263 4788 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53544ba9776f31487912a21c8121031d2cdf6b591498db55c511e03befba8381" Oct 10 17:33:37 crc kubenswrapper[4788]: I1010 17:33:37.919330 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/crc-debug-5b68h" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.254271 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58f8ca7c-88c0-45ed-b957-3e5000625007" path="/var/lib/kubelet/pods/58f8ca7c-88c0-45ed-b957-3e5000625007/volumes" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.559487 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7xjww/crc-debug-sbxpm"] Oct 10 17:33:38 crc kubenswrapper[4788]: E1010 17:33:38.559953 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58f8ca7c-88c0-45ed-b957-3e5000625007" containerName="container-00" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.559972 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="58f8ca7c-88c0-45ed-b957-3e5000625007" containerName="container-00" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.560221 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="58f8ca7c-88c0-45ed-b957-3e5000625007" containerName="container-00" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.560989 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.728966 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5zhm\" (UniqueName: \"kubernetes.io/projected/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-kube-api-access-b5zhm\") pod \"crc-debug-sbxpm\" (UID: \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\") " pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.729777 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-host\") pod \"crc-debug-sbxpm\" (UID: \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\") " pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.831191 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-host\") pod \"crc-debug-sbxpm\" (UID: \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\") " pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.831376 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-host\") pod \"crc-debug-sbxpm\" (UID: \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\") " pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.831932 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5zhm\" (UniqueName: \"kubernetes.io/projected/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-kube-api-access-b5zhm\") pod \"crc-debug-sbxpm\" (UID: \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\") " pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.861381 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5zhm\" (UniqueName: \"kubernetes.io/projected/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-kube-api-access-b5zhm\") pod \"crc-debug-sbxpm\" (UID: \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\") " pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.879732 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:38 crc kubenswrapper[4788]: W1010 17:33:38.917660 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8f243a6_e3ed_4ff8_9f9b_aa2e58201e8f.slice/crio-7e006809c6bcf35b90388f52fdd334432593a2ca9d498f906796828dba310f8e WatchSource:0}: Error finding container 7e006809c6bcf35b90388f52fdd334432593a2ca9d498f906796828dba310f8e: Status 404 returned error can't find the container with id 7e006809c6bcf35b90388f52fdd334432593a2ca9d498f906796828dba310f8e Oct 10 17:33:38 crc kubenswrapper[4788]: I1010 17:33:38.933948 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/crc-debug-sbxpm" event={"ID":"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f","Type":"ContainerStarted","Data":"7e006809c6bcf35b90388f52fdd334432593a2ca9d498f906796828dba310f8e"} Oct 10 17:33:39 crc kubenswrapper[4788]: I1010 17:33:39.960547 4788 generic.go:334] "Generic (PLEG): container finished" podID="d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f" containerID="1dc2cf0d73de950c1544e471eef101975c3851b01dfcdb2f82721a51a00467b3" exitCode=1 Oct 10 17:33:39 crc kubenswrapper[4788]: I1010 17:33:39.961213 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/crc-debug-sbxpm" event={"ID":"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f","Type":"ContainerDied","Data":"1dc2cf0d73de950c1544e471eef101975c3851b01dfcdb2f82721a51a00467b3"} Oct 10 17:33:40 crc kubenswrapper[4788]: I1010 17:33:40.021597 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7xjww/crc-debug-sbxpm"] Oct 10 17:33:40 crc kubenswrapper[4788]: I1010 17:33:40.036112 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7xjww/crc-debug-sbxpm"] Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.443679 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.495882 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-host\") pod \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\" (UID: \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\") " Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.495929 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-host" (OuterVolumeSpecName: "host") pod "d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f" (UID: "d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.496069 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5zhm\" (UniqueName: \"kubernetes.io/projected/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-kube-api-access-b5zhm\") pod \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\" (UID: \"d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f\") " Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.496661 4788 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-host\") on node \"crc\" DevicePath \"\"" Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.503479 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-kube-api-access-b5zhm" (OuterVolumeSpecName: "kube-api-access-b5zhm") pod "d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f" (UID: "d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f"). InnerVolumeSpecName "kube-api-access-b5zhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.599485 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5zhm\" (UniqueName: \"kubernetes.io/projected/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f-kube-api-access-b5zhm\") on node \"crc\" DevicePath \"\"" Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.988211 4788 scope.go:117] "RemoveContainer" containerID="1dc2cf0d73de950c1544e471eef101975c3851b01dfcdb2f82721a51a00467b3" Oct 10 17:33:41 crc kubenswrapper[4788]: I1010 17:33:41.988280 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/crc-debug-sbxpm" Oct 10 17:33:42 crc kubenswrapper[4788]: I1010 17:33:42.246834 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f" path="/var/lib/kubelet/pods/d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f/volumes" Oct 10 17:34:34 crc kubenswrapper[4788]: I1010 17:34:34.712993 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_766efb82-f5a8-44a0-b451-8ee5d7bdb508/init-config-reloader/0.log" Oct 10 17:34:34 crc kubenswrapper[4788]: I1010 17:34:34.997278 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_766efb82-f5a8-44a0-b451-8ee5d7bdb508/init-config-reloader/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.048034 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_766efb82-f5a8-44a0-b451-8ee5d7bdb508/config-reloader/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.094169 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_766efb82-f5a8-44a0-b451-8ee5d7bdb508/alertmanager/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.467743 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7b0a366f-f03c-4607-8e50-6df8bf30d683/aodh-api/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.507538 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7b0a366f-f03c-4607-8e50-6df8bf30d683/aodh-evaluator/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.632896 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7b0a366f-f03c-4607-8e50-6df8bf30d683/aodh-listener/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.698157 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7b0a366f-f03c-4607-8e50-6df8bf30d683/aodh-notifier/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.823066 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8578c7f644-zrr9p_77ccc4ea-3d0d-46b3-aa0f-ca0affce488c/barbican-api/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.901956 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8578c7f644-zrr9p_77ccc4ea-3d0d-46b3-aa0f-ca0affce488c/barbican-api-log/0.log" Oct 10 17:34:35 crc kubenswrapper[4788]: I1010 17:34:35.991435 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-656647f4cd-wp2dt_bbf57fa7-9537-46e6-9902-ea78b098412e/barbican-keystone-listener/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.135684 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-656647f4cd-wp2dt_bbf57fa7-9537-46e6-9902-ea78b098412e/barbican-keystone-listener-log/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.273559 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-86b75f4667-rhtqc_d8e57d00-cebc-4ea5-9033-c9009ed5c21f/barbican-worker/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.353312 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-86b75f4667-rhtqc_d8e57d00-cebc-4ea5-9033-c9009ed5c21f/barbican-worker-log/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.482536 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-xsj2v_171bf601-1ec1-4711-aacd-611751934a92/bootstrap-openstack-openstack-cell1/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.671842 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6d5dfc2e-d845-48b9-8741-d9972bbab228/ceilometer-central-agent/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.689563 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6d5dfc2e-d845-48b9-8741-d9972bbab228/ceilometer-notification-agent/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.785580 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6d5dfc2e-d845-48b9-8741-d9972bbab228/proxy-httpd/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.848846 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_6d5dfc2e-d845-48b9-8741-d9972bbab228/sg-core/0.log" Oct 10 17:34:36 crc kubenswrapper[4788]: I1010 17:34:36.973376 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-gc2fn_66598dac-a1d9-49ad-a8a6-f1d663d9907d/ceph-client-openstack-openstack-cell1/0.log" Oct 10 17:34:37 crc kubenswrapper[4788]: I1010 17:34:37.181436 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_15109044-050e-4ac9-b05e-8d5b80bf2f69/cinder-api-log/0.log" Oct 10 17:34:37 crc kubenswrapper[4788]: I1010 17:34:37.263761 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_15109044-050e-4ac9-b05e-8d5b80bf2f69/cinder-api/0.log" Oct 10 17:34:37 crc kubenswrapper[4788]: I1010 17:34:37.481626 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_f6ef5c29-c7b8-4181-ab0d-4725a545a8a7/probe/0.log" Oct 10 17:34:37 crc kubenswrapper[4788]: I1010 17:34:37.601499 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_f6ef5c29-c7b8-4181-ab0d-4725a545a8a7/cinder-backup/0.log" Oct 10 17:34:37 crc kubenswrapper[4788]: I1010 17:34:37.641722 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23d88b3c-33af-4075-8e30-1bca0b848ee3/cinder-scheduler/0.log" Oct 10 17:34:37 crc kubenswrapper[4788]: I1010 17:34:37.752179 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_23d88b3c-33af-4075-8e30-1bca0b848ee3/probe/0.log" Oct 10 17:34:37 crc kubenswrapper[4788]: I1010 17:34:37.873311 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_53564aa8-cbc4-4373-8330-ab61e819e9b7/cinder-volume/0.log" Oct 10 17:34:37 crc kubenswrapper[4788]: I1010 17:34:37.907838 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_53564aa8-cbc4-4373-8330-ab61e819e9b7/probe/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.042874 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-lf2lb_c64cc836-ad93-4dcf-a6f1-29e21606c64c/configure-network-openstack-openstack-cell1/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.151755 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-n5thk_d6bf08dd-da00-4e04-856f-8b6bff2eb1e6/configure-os-openstack-openstack-cell1/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.253674 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-84b74f9b95-7wj79_3c0ee6c9-eb8b-4054-a552-fccd55b68cfc/init/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.414566 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-84b74f9b95-7wj79_3c0ee6c9-eb8b-4054-a552-fccd55b68cfc/init/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.447482 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-84b74f9b95-7wj79_3c0ee6c9-eb8b-4054-a552-fccd55b68cfc/dnsmasq-dns/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.528822 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-88wwf_d4f46419-8e4b-408e-88aa-55d6bebb8ce9/download-cache-openstack-openstack-cell1/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.651498 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d65afb02-bfb8-4467-a722-1630581e302c/glance-httpd/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.703596 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d65afb02-bfb8-4467-a722-1630581e302c/glance-log/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.850827 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_31093023-af84-44fa-a6e7-a112b823b67c/glance-log/0.log" Oct 10 17:34:38 crc kubenswrapper[4788]: I1010 17:34:38.873616 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_31093023-af84-44fa-a6e7-a112b823b67c/glance-httpd/0.log" Oct 10 17:34:39 crc kubenswrapper[4788]: I1010 17:34:39.123004 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-f7fd69bfb-kdvkm_245f7946-b9af-40b8-b5f3-df13e6af95b8/heat-api/0.log" Oct 10 17:34:39 crc kubenswrapper[4788]: I1010 17:34:39.194794 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-dd78bbb65-x49zv_1e8995e7-5e84-4faf-b2d7-90d333960392/heat-cfnapi/0.log" Oct 10 17:34:39 crc kubenswrapper[4788]: I1010 17:34:39.290676 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7bcb6dbfb-zktx9_5c439913-7ebb-4ec3-921f-a7951566ef4b/heat-engine/0.log" Oct 10 17:34:39 crc kubenswrapper[4788]: I1010 17:34:39.442659 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-57bbb94c77-drsds_38bff86e-3dbc-45d4-b48e-236d63f80e30/horizon/0.log" Oct 10 17:34:39 crc kubenswrapper[4788]: I1010 17:34:39.567463 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-57bbb94c77-drsds_38bff86e-3dbc-45d4-b48e-236d63f80e30/horizon-log/0.log" Oct 10 17:34:39 crc kubenswrapper[4788]: I1010 17:34:39.633321 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-97dq9_6eff8747-d0bb-49d9-b45d-95ed5adf39a0/install-certs-openstack-openstack-cell1/0.log" Oct 10 17:34:39 crc kubenswrapper[4788]: I1010 17:34:39.783734 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-84z6g_2793e9af-a0b7-4dc7-91fc-44ea0f36366a/install-os-openstack-openstack-cell1/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.008198 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7444b6d9d6-vbfgf_325746a3-78bd-4fe7-9ff9-6003f15efc55/keystone-api/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.020890 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29335261-dvp6q_d252c223-b923-4821-a4ff-8c183277565d/keystone-cron/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.149849 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_1e3d4ec1-9e1e-4d22-90ce-69a7ee36b09c/kube-state-metrics/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.250684 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-47fvb_15644ee5-8d10-47e1-a124-a7df41949cda/libvirt-openstack-openstack-cell1/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.556909 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f6ed40a0-c7d7-4213-82f5-96fc17015781/manila-api/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.568324 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f6ed40a0-c7d7-4213-82f5-96fc17015781/manila-api-log/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.794892 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_732e8571-c5ae-4750-a39a-b51da05c0385/probe/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.843206 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_732e8571-c5ae-4750-a39a-b51da05c0385/manila-scheduler/0.log" Oct 10 17:34:40 crc kubenswrapper[4788]: I1010 17:34:40.991771 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_3499fc18-c5b2-4a16-9db0-3f74ba613448/probe/0.log" Oct 10 17:34:41 crc kubenswrapper[4788]: I1010 17:34:41.015040 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_3499fc18-c5b2-4a16-9db0-3f74ba613448/manila-share/0.log" Oct 10 17:34:41 crc kubenswrapper[4788]: I1010 17:34:41.067860 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-copy-data_8b0029af-55ed-45fc-b831-d115439a4591/adoption/0.log" Oct 10 17:34:41 crc kubenswrapper[4788]: I1010 17:34:41.447731 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-56d899f5c5-sdv5n_4057ec67-9bd6-43d0-941c-39d3373caa74/neutron-api/0.log" Oct 10 17:34:41 crc kubenswrapper[4788]: I1010 17:34:41.546939 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-56d899f5c5-sdv5n_4057ec67-9bd6-43d0-941c-39d3373caa74/neutron-httpd/0.log" Oct 10 17:34:41 crc kubenswrapper[4788]: I1010 17:34:41.797851 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-29wxw_6bc23201-2e25-46eb-a733-4585e2356f16/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 10 17:34:42 crc kubenswrapper[4788]: I1010 17:34:42.026626 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-mgbkw_83f651a1-892d-4422-964b-93c92ddfa185/neutron-metadata-openstack-openstack-cell1/0.log" Oct 10 17:34:42 crc kubenswrapper[4788]: I1010 17:34:42.334974 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-6xzwg_0f131e4c-500d-4914-9b97-869a2765d81e/neutron-sriov-openstack-openstack-cell1/0.log" Oct 10 17:34:42 crc kubenswrapper[4788]: I1010 17:34:42.642259 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2029803d-4cea-46d6-af6c-7a0b17a4733c/nova-api-api/0.log" Oct 10 17:34:42 crc kubenswrapper[4788]: I1010 17:34:42.823742 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2029803d-4cea-46d6-af6c-7a0b17a4733c/nova-api-log/0.log" Oct 10 17:34:42 crc kubenswrapper[4788]: I1010 17:34:42.948412 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_370a716b-2910-4149-9722-fb383dab7e4c/nova-cell0-conductor-conductor/0.log" Oct 10 17:34:43 crc kubenswrapper[4788]: I1010 17:34:43.184701 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_d37aad02-67ff-4f8f-af27-55d1acaed2e7/nova-cell1-conductor-conductor/0.log" Oct 10 17:34:43 crc kubenswrapper[4788]: I1010 17:34:43.377054 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2c017349-955a-4b6a-9ab3-b689d0316e6f/nova-cell1-novncproxy-novncproxy/0.log" Oct 10 17:34:43 crc kubenswrapper[4788]: I1010 17:34:43.623638 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celljbmbk_7d34aa65-68bf-476c-9b37-84a544e0eb4b/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 10 17:34:43 crc kubenswrapper[4788]: I1010 17:34:43.898701 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-cjc4m_e1c3f042-7a1a-4a21-accd-636a15e88010/nova-cell1-openstack-openstack-cell1/0.log" Oct 10 17:34:43 crc kubenswrapper[4788]: I1010 17:34:43.977414 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_19ed4b86-3570-4e57-a9f8-44e17a81beda/nova-metadata-log/0.log" Oct 10 17:34:44 crc kubenswrapper[4788]: I1010 17:34:44.096449 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_19ed4b86-3570-4e57-a9f8-44e17a81beda/nova-metadata-metadata/0.log" Oct 10 17:34:44 crc kubenswrapper[4788]: I1010 17:34:44.360841 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_6b5b6539-1514-46d5-8bfe-bfc92da4c9a0/nova-scheduler-scheduler/0.log" Oct 10 17:34:44 crc kubenswrapper[4788]: I1010 17:34:44.383940 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_230985d2-0cc5-499d-ad3a-dd37be0b450b/memcached/0.log" Oct 10 17:34:44 crc kubenswrapper[4788]: I1010 17:34:44.448643 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7ffd598478-ntz5f_18a51038-256b-4638-9ca3-44a8c87718ab/init/0.log" Oct 10 17:34:44 crc kubenswrapper[4788]: I1010 17:34:44.617233 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7ffd598478-ntz5f_18a51038-256b-4638-9ca3-44a8c87718ab/init/0.log" Oct 10 17:34:44 crc kubenswrapper[4788]: I1010 17:34:44.738114 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7ffd598478-ntz5f_18a51038-256b-4638-9ca3-44a8c87718ab/octavia-api-provider-agent/0.log" Oct 10 17:34:44 crc kubenswrapper[4788]: I1010 17:34:44.819891 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7ffd598478-ntz5f_18a51038-256b-4638-9ca3-44a8c87718ab/octavia-api/0.log" Oct 10 17:34:44 crc kubenswrapper[4788]: I1010 17:34:44.919408 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-dlvvq_6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1/init/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.115064 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-dlvvq_6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1/init/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.216176 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-dlvvq_6dd3f0aa-e9c2-4d5b-a2d2-f02df5fe23c1/octavia-healthmanager/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.216214 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-9n42v_e74a446b-ff90-447e-82c4-a61d927dc0a3/init/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.348219 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-9n42v_e74a446b-ff90-447e-82c4-a61d927dc0a3/init/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.406640 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-jttq6_9efa8985-a060-40c9-ba49-47e7d866c630/init/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.451684 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-9n42v_e74a446b-ff90-447e-82c4-a61d927dc0a3/octavia-housekeeping/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.611897 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-jttq6_9efa8985-a060-40c9-ba49-47e7d866c630/init/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.634442 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-jttq6_9efa8985-a060-40c9-ba49-47e7d866c630/octavia-rsyslog/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.727085 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-gbd4d_89b1209f-3797-42be-9f07-76b7ec120daf/init/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.888849 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-gbd4d_89b1209f-3797-42be-9f07-76b7ec120daf/init/0.log" Oct 10 17:34:45 crc kubenswrapper[4788]: I1010 17:34:45.969824 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a22008ee-3a48-4a02-9301-506b7133a8a5/mysql-bootstrap/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.031747 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-gbd4d_89b1209f-3797-42be-9f07-76b7ec120daf/octavia-worker/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.252988 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a22008ee-3a48-4a02-9301-506b7133a8a5/mysql-bootstrap/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.297682 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1690fbb-be06-482f-9c94-4d436d9b73b1/mysql-bootstrap/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.302646 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a22008ee-3a48-4a02-9301-506b7133a8a5/galera/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.479983 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1690fbb-be06-482f-9c94-4d436d9b73b1/mysql-bootstrap/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.502670 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1690fbb-be06-482f-9c94-4d436d9b73b1/galera/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.524795 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_8133c683-728c-4758-9c20-6eddde277d95/openstackclient/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.695022 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5k2fn_b7674601-b383-40b8-80ad-0864b4114d70/ovn-controller/0.log" Oct 10 17:34:46 crc kubenswrapper[4788]: I1010 17:34:46.713280 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-pntng_4e9e9129-a513-4255-9561-6a5f268193f2/openstack-network-exporter/0.log" Oct 10 17:34:47 crc kubenswrapper[4788]: I1010 17:34:47.140329 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t7sxw_a2b7ee70-2711-4155-a154-1fe0b96984e5/ovsdb-server-init/0.log" Oct 10 17:34:47 crc kubenswrapper[4788]: I1010 17:34:47.498198 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t7sxw_a2b7ee70-2711-4155-a154-1fe0b96984e5/ovsdb-server-init/0.log" Oct 10 17:34:47 crc kubenswrapper[4788]: I1010 17:34:47.502426 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t7sxw_a2b7ee70-2711-4155-a154-1fe0b96984e5/ovsdb-server/0.log" Oct 10 17:34:47 crc kubenswrapper[4788]: I1010 17:34:47.505833 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t7sxw_a2b7ee70-2711-4155-a154-1fe0b96984e5/ovs-vswitchd/0.log" Oct 10 17:34:47 crc kubenswrapper[4788]: I1010 17:34:47.656840 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-copy-data_dfa6f880-c8b3-49a4-9397-022761316eb3/adoption/0.log" Oct 10 17:34:47 crc kubenswrapper[4788]: I1010 17:34:47.728094 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_511d54ed-e20b-4439-9a8f-ab1f738fb5c7/openstack-network-exporter/0.log" Oct 10 17:34:47 crc kubenswrapper[4788]: I1010 17:34:47.751636 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_511d54ed-e20b-4439-9a8f-ab1f738fb5c7/ovn-northd/0.log" Oct 10 17:34:47 crc kubenswrapper[4788]: I1010 17:34:47.961131 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-qbrs7_bc21acf4-40c2-4cab-998e-5561eb853813/ovn-openstack-openstack-cell1/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.004039 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6d7c4590-0091-4af7-b1cf-98c2afbf1cbb/openstack-network-exporter/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.105057 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6d7c4590-0091-4af7-b1cf-98c2afbf1cbb/ovsdbserver-nb/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.220586 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2/ovsdbserver-nb/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.230561 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_237d4343-a4d1-42c4-bc8b-1e49ebd7f6a2/openstack-network-exporter/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.469830 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_6e5012dd-51b6-4db7-8371-b71f2cc46ec3/ovsdbserver-nb/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.484111 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_6e5012dd-51b6-4db7-8371-b71f2cc46ec3/openstack-network-exporter/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.632946 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_98ebdaa9-c26a-406d-b56f-68c4a9cf8da3/ovsdbserver-sb/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.640409 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_98ebdaa9-c26a-406d-b56f-68c4a9cf8da3/openstack-network-exporter/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.770160 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_892d027b-e0f3-420b-a4db-09926ccbf288/openstack-network-exporter/0.log" Oct 10 17:34:48 crc kubenswrapper[4788]: I1010 17:34:48.853571 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_892d027b-e0f3-420b-a4db-09926ccbf288/ovsdbserver-sb/0.log" Oct 10 17:34:49 crc kubenswrapper[4788]: I1010 17:34:49.562326 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_5e313987-6d14-4ea1-8ba6-616a31065713/openstack-network-exporter/0.log" Oct 10 17:34:49 crc kubenswrapper[4788]: I1010 17:34:49.591087 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_5e313987-6d14-4ea1-8ba6-616a31065713/ovsdbserver-sb/0.log" Oct 10 17:34:49 crc kubenswrapper[4788]: I1010 17:34:49.634052 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-75f45f9948-dj9bl_cade6027-5453-4492-919e-62a481d01439/placement-api/0.log" Oct 10 17:34:49 crc kubenswrapper[4788]: I1010 17:34:49.794720 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-75f45f9948-dj9bl_cade6027-5453-4492-919e-62a481d01439/placement-log/0.log" Oct 10 17:34:49 crc kubenswrapper[4788]: I1010 17:34:49.824317 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cwfxzs_17d65674-bacd-4214-a197-2b78b4f9f45d/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.001737 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2fd93ba6-6366-4f85-b5f3-cef3ed588a50/init-config-reloader/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.133430 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2fd93ba6-6366-4f85-b5f3-cef3ed588a50/init-config-reloader/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.152664 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2fd93ba6-6366-4f85-b5f3-cef3ed588a50/config-reloader/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.158027 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2fd93ba6-6366-4f85-b5f3-cef3ed588a50/prometheus/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.187743 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2fd93ba6-6366-4f85-b5f3-cef3ed588a50/thanos-sidecar/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.351217 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f0339776-eb9a-41e2-aea6-7b16469ea1a9/setup-container/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.535561 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f0339776-eb9a-41e2-aea6-7b16469ea1a9/setup-container/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.556052 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f0339776-eb9a-41e2-aea6-7b16469ea1a9/rabbitmq/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.635690 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec/setup-container/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.791466 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec/setup-container/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.816497 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cc92b5ff-1e72-45e4-a4ee-d0865a5cd3ec/rabbitmq/0.log" Oct 10 17:34:50 crc kubenswrapper[4788]: I1010 17:34:50.853553 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-d5bk5_38fe23d3-2d05-463f-9bea-67a8fe0cb377/reboot-os-openstack-openstack-cell1/0.log" Oct 10 17:34:51 crc kubenswrapper[4788]: I1010 17:34:51.017177 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-dnvvt_480d7aad-e180-4e43-8179-b6e1d63dc616/run-os-openstack-openstack-cell1/0.log" Oct 10 17:34:51 crc kubenswrapper[4788]: I1010 17:34:51.082545 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-rz6fx_1eb686cf-9ffe-4b1d-ad32-3800206d3c7d/ssh-known-hosts-openstack/0.log" Oct 10 17:34:51 crc kubenswrapper[4788]: I1010 17:34:51.292703 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-2ddfx_e4a0577d-9b0e-44ae-9b49-14374f508c62/telemetry-openstack-openstack-cell1/0.log" Oct 10 17:34:51 crc kubenswrapper[4788]: I1010 17:34:51.705681 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-5cw9q_6dbab995-d2d3-42cf-8f11-1484c7d00f4c/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 10 17:34:51 crc kubenswrapper[4788]: I1010 17:34:51.733874 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-4dtfv_bcdfb941-0c8e-4c5d-8043-19c13616166a/validate-network-openstack-openstack-cell1/0.log" Oct 10 17:34:59 crc kubenswrapper[4788]: I1010 17:34:59.406457 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:34:59 crc kubenswrapper[4788]: I1010 17:34:59.407216 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.724996 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pxx9s"] Oct 10 17:35:20 crc kubenswrapper[4788]: E1010 17:35:20.726249 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f" containerName="container-00" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.726268 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f" containerName="container-00" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.726662 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f243a6-e3ed-4ff8-9f9b-aa2e58201e8f" containerName="container-00" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.729562 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.739406 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxx9s"] Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.820857 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww4nz\" (UniqueName: \"kubernetes.io/projected/ba1217ab-5440-46ca-8b91-b263a85361b9-kube-api-access-ww4nz\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.821167 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-catalog-content\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.821613 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-utilities\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.923492 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-utilities\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.923587 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww4nz\" (UniqueName: \"kubernetes.io/projected/ba1217ab-5440-46ca-8b91-b263a85361b9-kube-api-access-ww4nz\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.923656 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-catalog-content\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.924126 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-catalog-content\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.924133 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-utilities\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:20 crc kubenswrapper[4788]: I1010 17:35:20.946183 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww4nz\" (UniqueName: \"kubernetes.io/projected/ba1217ab-5440-46ca-8b91-b263a85361b9-kube-api-access-ww4nz\") pod \"redhat-operators-pxx9s\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:21 crc kubenswrapper[4788]: I1010 17:35:21.072124 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:21 crc kubenswrapper[4788]: I1010 17:35:21.616118 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxx9s"] Oct 10 17:35:22 crc kubenswrapper[4788]: I1010 17:35:22.062171 4788 generic.go:334] "Generic (PLEG): container finished" podID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerID="1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35" exitCode=0 Oct 10 17:35:22 crc kubenswrapper[4788]: I1010 17:35:22.062390 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxx9s" event={"ID":"ba1217ab-5440-46ca-8b91-b263a85361b9","Type":"ContainerDied","Data":"1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35"} Oct 10 17:35:22 crc kubenswrapper[4788]: I1010 17:35:22.062604 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxx9s" event={"ID":"ba1217ab-5440-46ca-8b91-b263a85361b9","Type":"ContainerStarted","Data":"3f602177b2fa5d616bab9e299b27fa221afa1bf62b5547cf04947854c575474b"} Oct 10 17:35:22 crc kubenswrapper[4788]: I1010 17:35:22.064995 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 17:35:23 crc kubenswrapper[4788]: I1010 17:35:23.073214 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxx9s" event={"ID":"ba1217ab-5440-46ca-8b91-b263a85361b9","Type":"ContainerStarted","Data":"04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5"} Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.116413 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7h9fc"] Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.120306 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.132361 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h9fc"] Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.261538 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-utilities\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.261639 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvbpk\" (UniqueName: \"kubernetes.io/projected/70bfd12c-3169-4c20-ae69-41f5421c3071-kube-api-access-lvbpk\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.261924 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-catalog-content\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.363102 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-utilities\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.363187 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvbpk\" (UniqueName: \"kubernetes.io/projected/70bfd12c-3169-4c20-ae69-41f5421c3071-kube-api-access-lvbpk\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.363270 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-catalog-content\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.363573 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-utilities\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.363646 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-catalog-content\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.387452 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvbpk\" (UniqueName: \"kubernetes.io/projected/70bfd12c-3169-4c20-ae69-41f5421c3071-kube-api-access-lvbpk\") pod \"redhat-marketplace-7h9fc\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.459019 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:26 crc kubenswrapper[4788]: I1010 17:35:26.964317 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h9fc"] Oct 10 17:35:26 crc kubenswrapper[4788]: W1010 17:35:26.966232 4788 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70bfd12c_3169_4c20_ae69_41f5421c3071.slice/crio-a6a7f9b856ab1f32bcbecbb3f30cb1ec37936189820d4ca639dacc43fc7f13ff WatchSource:0}: Error finding container a6a7f9b856ab1f32bcbecbb3f30cb1ec37936189820d4ca639dacc43fc7f13ff: Status 404 returned error can't find the container with id a6a7f9b856ab1f32bcbecbb3f30cb1ec37936189820d4ca639dacc43fc7f13ff Oct 10 17:35:27 crc kubenswrapper[4788]: I1010 17:35:27.120001 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h9fc" event={"ID":"70bfd12c-3169-4c20-ae69-41f5421c3071","Type":"ContainerStarted","Data":"a6a7f9b856ab1f32bcbecbb3f30cb1ec37936189820d4ca639dacc43fc7f13ff"} Oct 10 17:35:27 crc kubenswrapper[4788]: I1010 17:35:27.126385 4788 generic.go:334] "Generic (PLEG): container finished" podID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerID="04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5" exitCode=0 Oct 10 17:35:27 crc kubenswrapper[4788]: I1010 17:35:27.126446 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxx9s" event={"ID":"ba1217ab-5440-46ca-8b91-b263a85361b9","Type":"ContainerDied","Data":"04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5"} Oct 10 17:35:28 crc kubenswrapper[4788]: I1010 17:35:28.141711 4788 generic.go:334] "Generic (PLEG): container finished" podID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerID="305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea" exitCode=0 Oct 10 17:35:28 crc kubenswrapper[4788]: I1010 17:35:28.141783 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h9fc" event={"ID":"70bfd12c-3169-4c20-ae69-41f5421c3071","Type":"ContainerDied","Data":"305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea"} Oct 10 17:35:28 crc kubenswrapper[4788]: I1010 17:35:28.146228 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxx9s" event={"ID":"ba1217ab-5440-46ca-8b91-b263a85361b9","Type":"ContainerStarted","Data":"358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45"} Oct 10 17:35:28 crc kubenswrapper[4788]: I1010 17:35:28.196344 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pxx9s" podStartSLOduration=2.726562021 podStartE2EDuration="8.196319047s" podCreationTimestamp="2025-10-10 17:35:20 +0000 UTC" firstStartedPulling="2025-10-10 17:35:22.064749839 +0000 UTC m=+10224.514465387" lastFinishedPulling="2025-10-10 17:35:27.534506835 +0000 UTC m=+10229.984222413" observedRunningTime="2025-10-10 17:35:28.19529403 +0000 UTC m=+10230.645009608" watchObservedRunningTime="2025-10-10 17:35:28.196319047 +0000 UTC m=+10230.646034605" Oct 10 17:35:29 crc kubenswrapper[4788]: I1010 17:35:29.406494 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:35:29 crc kubenswrapper[4788]: I1010 17:35:29.407362 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:35:30 crc kubenswrapper[4788]: I1010 17:35:30.179286 4788 generic.go:334] "Generic (PLEG): container finished" podID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerID="b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191" exitCode=0 Oct 10 17:35:30 crc kubenswrapper[4788]: I1010 17:35:30.179393 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h9fc" event={"ID":"70bfd12c-3169-4c20-ae69-41f5421c3071","Type":"ContainerDied","Data":"b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191"} Oct 10 17:35:31 crc kubenswrapper[4788]: I1010 17:35:31.072369 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:31 crc kubenswrapper[4788]: I1010 17:35:31.072791 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:31 crc kubenswrapper[4788]: I1010 17:35:31.194256 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h9fc" event={"ID":"70bfd12c-3169-4c20-ae69-41f5421c3071","Type":"ContainerStarted","Data":"b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181"} Oct 10 17:35:31 crc kubenswrapper[4788]: I1010 17:35:31.225072 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7h9fc" podStartSLOduration=2.708842037 podStartE2EDuration="5.225043345s" podCreationTimestamp="2025-10-10 17:35:26 +0000 UTC" firstStartedPulling="2025-10-10 17:35:28.144242129 +0000 UTC m=+10230.593957687" lastFinishedPulling="2025-10-10 17:35:30.660443447 +0000 UTC m=+10233.110158995" observedRunningTime="2025-10-10 17:35:31.222391935 +0000 UTC m=+10233.672107483" watchObservedRunningTime="2025-10-10 17:35:31.225043345 +0000 UTC m=+10233.674758893" Oct 10 17:35:32 crc kubenswrapper[4788]: I1010 17:35:32.128485 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pxx9s" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="registry-server" probeResult="failure" output=< Oct 10 17:35:32 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 17:35:32 crc kubenswrapper[4788]: > Oct 10 17:35:36 crc kubenswrapper[4788]: I1010 17:35:36.459392 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:36 crc kubenswrapper[4788]: I1010 17:35:36.460189 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:36 crc kubenswrapper[4788]: I1010 17:35:36.649624 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:37 crc kubenswrapper[4788]: I1010 17:35:37.360389 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:37 crc kubenswrapper[4788]: I1010 17:35:37.426355 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h9fc"] Oct 10 17:35:39 crc kubenswrapper[4788]: I1010 17:35:39.314432 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7h9fc" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerName="registry-server" containerID="cri-o://b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181" gracePeriod=2 Oct 10 17:35:39 crc kubenswrapper[4788]: I1010 17:35:39.935408 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.036721 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-catalog-content\") pod \"70bfd12c-3169-4c20-ae69-41f5421c3071\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.036899 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-utilities\") pod \"70bfd12c-3169-4c20-ae69-41f5421c3071\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.037046 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvbpk\" (UniqueName: \"kubernetes.io/projected/70bfd12c-3169-4c20-ae69-41f5421c3071-kube-api-access-lvbpk\") pod \"70bfd12c-3169-4c20-ae69-41f5421c3071\" (UID: \"70bfd12c-3169-4c20-ae69-41f5421c3071\") " Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.037647 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-utilities" (OuterVolumeSpecName: "utilities") pod "70bfd12c-3169-4c20-ae69-41f5421c3071" (UID: "70bfd12c-3169-4c20-ae69-41f5421c3071"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.038010 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.043365 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70bfd12c-3169-4c20-ae69-41f5421c3071-kube-api-access-lvbpk" (OuterVolumeSpecName: "kube-api-access-lvbpk") pod "70bfd12c-3169-4c20-ae69-41f5421c3071" (UID: "70bfd12c-3169-4c20-ae69-41f5421c3071"). InnerVolumeSpecName "kube-api-access-lvbpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.054040 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70bfd12c-3169-4c20-ae69-41f5421c3071" (UID: "70bfd12c-3169-4c20-ae69-41f5421c3071"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.141822 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70bfd12c-3169-4c20-ae69-41f5421c3071-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.141861 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvbpk\" (UniqueName: \"kubernetes.io/projected/70bfd12c-3169-4c20-ae69-41f5421c3071-kube-api-access-lvbpk\") on node \"crc\" DevicePath \"\"" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.326973 4788 generic.go:334] "Generic (PLEG): container finished" podID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerID="b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181" exitCode=0 Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.327059 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h9fc" event={"ID":"70bfd12c-3169-4c20-ae69-41f5421c3071","Type":"ContainerDied","Data":"b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181"} Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.327100 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7h9fc" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.327175 4788 scope.go:117] "RemoveContainer" containerID="b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.327182 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h9fc" event={"ID":"70bfd12c-3169-4c20-ae69-41f5421c3071","Type":"ContainerDied","Data":"a6a7f9b856ab1f32bcbecbb3f30cb1ec37936189820d4ca639dacc43fc7f13ff"} Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.361377 4788 scope.go:117] "RemoveContainer" containerID="b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.364039 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h9fc"] Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.380243 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h9fc"] Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.397105 4788 scope.go:117] "RemoveContainer" containerID="305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.434419 4788 scope.go:117] "RemoveContainer" containerID="b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181" Oct 10 17:35:40 crc kubenswrapper[4788]: E1010 17:35:40.435007 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181\": container with ID starting with b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181 not found: ID does not exist" containerID="b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.435081 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181"} err="failed to get container status \"b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181\": rpc error: code = NotFound desc = could not find container \"b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181\": container with ID starting with b1a89da5560fa18297cd08692b58fa5dce6cb96fa16f1b6d24d0449c3266b181 not found: ID does not exist" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.435184 4788 scope.go:117] "RemoveContainer" containerID="b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191" Oct 10 17:35:40 crc kubenswrapper[4788]: E1010 17:35:40.435557 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191\": container with ID starting with b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191 not found: ID does not exist" containerID="b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.435637 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191"} err="failed to get container status \"b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191\": rpc error: code = NotFound desc = could not find container \"b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191\": container with ID starting with b48552d34ab033f123f37e664ea630d159d3d27ad30f12003359d917a5756191 not found: ID does not exist" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.435681 4788 scope.go:117] "RemoveContainer" containerID="305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea" Oct 10 17:35:40 crc kubenswrapper[4788]: E1010 17:35:40.436397 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea\": container with ID starting with 305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea not found: ID does not exist" containerID="305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea" Oct 10 17:35:40 crc kubenswrapper[4788]: I1010 17:35:40.436450 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea"} err="failed to get container status \"305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea\": rpc error: code = NotFound desc = could not find container \"305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea\": container with ID starting with 305a98525cc98d22e8e5f858bb423a8358a400cca1c241a8f5e776d9dbbfc9ea not found: ID does not exist" Oct 10 17:35:42 crc kubenswrapper[4788]: I1010 17:35:42.123590 4788 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pxx9s" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="registry-server" probeResult="failure" output=< Oct 10 17:35:42 crc kubenswrapper[4788]: timeout: failed to connect service ":50051" within 1s Oct 10 17:35:42 crc kubenswrapper[4788]: > Oct 10 17:35:42 crc kubenswrapper[4788]: I1010 17:35:42.250183 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" path="/var/lib/kubelet/pods/70bfd12c-3169-4c20-ae69-41f5421c3071/volumes" Oct 10 17:35:51 crc kubenswrapper[4788]: I1010 17:35:51.159455 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:51 crc kubenswrapper[4788]: I1010 17:35:51.225115 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:51 crc kubenswrapper[4788]: I1010 17:35:51.924236 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxx9s"] Oct 10 17:35:52 crc kubenswrapper[4788]: I1010 17:35:52.517072 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pxx9s" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="registry-server" containerID="cri-o://358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45" gracePeriod=2 Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.055770 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.168449 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww4nz\" (UniqueName: \"kubernetes.io/projected/ba1217ab-5440-46ca-8b91-b263a85361b9-kube-api-access-ww4nz\") pod \"ba1217ab-5440-46ca-8b91-b263a85361b9\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.168593 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-utilities\") pod \"ba1217ab-5440-46ca-8b91-b263a85361b9\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.168750 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-catalog-content\") pod \"ba1217ab-5440-46ca-8b91-b263a85361b9\" (UID: \"ba1217ab-5440-46ca-8b91-b263a85361b9\") " Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.169691 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-utilities" (OuterVolumeSpecName: "utilities") pod "ba1217ab-5440-46ca-8b91-b263a85361b9" (UID: "ba1217ab-5440-46ca-8b91-b263a85361b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.182737 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba1217ab-5440-46ca-8b91-b263a85361b9-kube-api-access-ww4nz" (OuterVolumeSpecName: "kube-api-access-ww4nz") pod "ba1217ab-5440-46ca-8b91-b263a85361b9" (UID: "ba1217ab-5440-46ca-8b91-b263a85361b9"). InnerVolumeSpecName "kube-api-access-ww4nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.255180 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba1217ab-5440-46ca-8b91-b263a85361b9" (UID: "ba1217ab-5440-46ca-8b91-b263a85361b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.271890 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.271930 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww4nz\" (UniqueName: \"kubernetes.io/projected/ba1217ab-5440-46ca-8b91-b263a85361b9-kube-api-access-ww4nz\") on node \"crc\" DevicePath \"\"" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.271947 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1217ab-5440-46ca-8b91-b263a85361b9-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.535692 4788 generic.go:334] "Generic (PLEG): container finished" podID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerID="358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45" exitCode=0 Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.535776 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxx9s" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.535790 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxx9s" event={"ID":"ba1217ab-5440-46ca-8b91-b263a85361b9","Type":"ContainerDied","Data":"358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45"} Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.536651 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxx9s" event={"ID":"ba1217ab-5440-46ca-8b91-b263a85361b9","Type":"ContainerDied","Data":"3f602177b2fa5d616bab9e299b27fa221afa1bf62b5547cf04947854c575474b"} Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.536702 4788 scope.go:117] "RemoveContainer" containerID="358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.556357 4788 scope.go:117] "RemoveContainer" containerID="04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.577922 4788 scope.go:117] "RemoveContainer" containerID="1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.649093 4788 scope.go:117] "RemoveContainer" containerID="358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45" Oct 10 17:35:53 crc kubenswrapper[4788]: E1010 17:35:53.649694 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45\": container with ID starting with 358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45 not found: ID does not exist" containerID="358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.649765 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45"} err="failed to get container status \"358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45\": rpc error: code = NotFound desc = could not find container \"358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45\": container with ID starting with 358cc5b7cea962b62062839f0c965e038653dea149a7bad9369919f88357dc45 not found: ID does not exist" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.649815 4788 scope.go:117] "RemoveContainer" containerID="04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5" Oct 10 17:35:53 crc kubenswrapper[4788]: E1010 17:35:53.650521 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5\": container with ID starting with 04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5 not found: ID does not exist" containerID="04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.650558 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5"} err="failed to get container status \"04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5\": rpc error: code = NotFound desc = could not find container \"04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5\": container with ID starting with 04a9f8a4235d09ff6d6ed2b249630061ccfd039039f572c7d593d3a26e2fd1f5 not found: ID does not exist" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.650584 4788 scope.go:117] "RemoveContainer" containerID="1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35" Oct 10 17:35:53 crc kubenswrapper[4788]: E1010 17:35:53.650948 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35\": container with ID starting with 1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35 not found: ID does not exist" containerID="1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.650997 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35"} err="failed to get container status \"1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35\": rpc error: code = NotFound desc = could not find container \"1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35\": container with ID starting with 1099f839ab79e7c657ed5a1963615f6addd64ea276bdf95234da970adc643b35 not found: ID does not exist" Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.658691 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxx9s"] Oct 10 17:35:53 crc kubenswrapper[4788]: I1010 17:35:53.668869 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pxx9s"] Oct 10 17:35:54 crc kubenswrapper[4788]: I1010 17:35:54.264298 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" path="/var/lib/kubelet/pods/ba1217ab-5440-46ca-8b91-b263a85361b9/volumes" Oct 10 17:35:59 crc kubenswrapper[4788]: I1010 17:35:59.406441 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:35:59 crc kubenswrapper[4788]: I1010 17:35:59.407609 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:35:59 crc kubenswrapper[4788]: I1010 17:35:59.407694 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 17:35:59 crc kubenswrapper[4788]: I1010 17:35:59.409209 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93046d31ffbebfcb70b32194820289c869f35af289dc70cd1e47ca5fb7e4b2bf"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 17:35:59 crc kubenswrapper[4788]: I1010 17:35:59.409335 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://93046d31ffbebfcb70b32194820289c869f35af289dc70cd1e47ca5fb7e4b2bf" gracePeriod=600 Oct 10 17:35:59 crc kubenswrapper[4788]: I1010 17:35:59.665614 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="93046d31ffbebfcb70b32194820289c869f35af289dc70cd1e47ca5fb7e4b2bf" exitCode=0 Oct 10 17:35:59 crc kubenswrapper[4788]: I1010 17:35:59.665679 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"93046d31ffbebfcb70b32194820289c869f35af289dc70cd1e47ca5fb7e4b2bf"} Oct 10 17:35:59 crc kubenswrapper[4788]: I1010 17:35:59.666110 4788 scope.go:117] "RemoveContainer" containerID="084af1ec7499aca565403bdef2663aa5d137d67936011f5330943d04dfacab1b" Oct 10 17:36:00 crc kubenswrapper[4788]: I1010 17:36:00.693406 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerStarted","Data":"ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac"} Oct 10 17:36:32 crc kubenswrapper[4788]: I1010 17:36:32.590816 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-k5ls7_1d47d678-34c1-41d7-a8cf-6f13199cc741/kube-rbac-proxy/0.log" Oct 10 17:36:32 crc kubenswrapper[4788]: I1010 17:36:32.691017 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-k5ls7_1d47d678-34c1-41d7-a8cf-6f13199cc741/manager/0.log" Oct 10 17:36:32 crc kubenswrapper[4788]: I1010 17:36:32.799022 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-9pbsq_2f9d641b-50d5-4dd9-baec-3c22d23273e1/kube-rbac-proxy/0.log" Oct 10 17:36:32 crc kubenswrapper[4788]: I1010 17:36:32.918115 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-9pbsq_2f9d641b-50d5-4dd9-baec-3c22d23273e1/manager/0.log" Oct 10 17:36:32 crc kubenswrapper[4788]: I1010 17:36:32.936931 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ckw79_84cbe7f0-4f13-468c-af71-022960cd0973/kube-rbac-proxy/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.007201 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ckw79_84cbe7f0-4f13-468c-af71-022960cd0973/manager/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.128342 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp_a7f6140b-914c-4de9-9ab3-11cb6fd83225/util/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.291126 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp_a7f6140b-914c-4de9-9ab3-11cb6fd83225/util/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.297492 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp_a7f6140b-914c-4de9-9ab3-11cb6fd83225/pull/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.335956 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp_a7f6140b-914c-4de9-9ab3-11cb6fd83225/pull/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.512313 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp_a7f6140b-914c-4de9-9ab3-11cb6fd83225/util/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.575105 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp_a7f6140b-914c-4de9-9ab3-11cb6fd83225/pull/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.601660 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fed058dffaf4d25ec897b5982a5c8e43d58f83d0db68dd9f92bb4ec76djzmkp_a7f6140b-914c-4de9-9ab3-11cb6fd83225/extract/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.705681 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-5v88l_8d40607c-b865-4722-93db-94f3fb0b2f42/kube-rbac-proxy/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.823026 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-929g7_9c8f6b95-2440-4fd6-b0a9-da1b17facf7e/kube-rbac-proxy/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.876442 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-5v88l_8d40607c-b865-4722-93db-94f3fb0b2f42/manager/0.log" Oct 10 17:36:33 crc kubenswrapper[4788]: I1010 17:36:33.972449 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-929g7_9c8f6b95-2440-4fd6-b0a9-da1b17facf7e/manager/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.060881 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-j4xpv_45ccc74a-b207-4512-86fd-07e0495d1378/kube-rbac-proxy/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.077112 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-j4xpv_45ccc74a-b207-4512-86fd-07e0495d1378/manager/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.276732 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-l678z_4e656fbd-3ed0-4434-9264-5a08a20607aa/kube-rbac-proxy/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.427542 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-5hstt_f3166185-7572-4637-b2c6-3b81a1cd79af/kube-rbac-proxy/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.539097 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-l678z_4e656fbd-3ed0-4434-9264-5a08a20607aa/manager/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.560555 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-5hstt_f3166185-7572-4637-b2c6-3b81a1cd79af/manager/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.629349 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-4spql_67694f95-9859-40b3-ae66-65cfdc1d17f5/kube-rbac-proxy/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.798416 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-8l7ct_aae7a82d-e804-4f31-a8be-cc52317be3f1/kube-rbac-proxy/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.809223 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-4spql_67694f95-9859-40b3-ae66-65cfdc1d17f5/manager/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.892206 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-8l7ct_aae7a82d-e804-4f31-a8be-cc52317be3f1/manager/0.log" Oct 10 17:36:34 crc kubenswrapper[4788]: I1010 17:36:34.987651 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-pkc5n_682f41df-b648-476f-8a3e-9475501bf290/kube-rbac-proxy/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.061216 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-pkc5n_682f41df-b648-476f-8a3e-9475501bf290/manager/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.204438 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-mvw9d_41d202a5-eddf-4ee7-9b9a-9edf7905d51f/kube-rbac-proxy/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.352118 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-mvw9d_41d202a5-eddf-4ee7-9b9a-9edf7905d51f/manager/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.383175 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-2cwz7_6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0/kube-rbac-proxy/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.633037 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-x2mn7_6791183a-bdf7-4ce2-b888-cd7e1b70ab80/kube-rbac-proxy/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.685439 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-2cwz7_6ddf56dc-44b6-49a7-9bf6-91362cdd3ea0/manager/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.755455 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-x2mn7_6791183a-bdf7-4ce2-b888-cd7e1b70ab80/manager/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.834132 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz_07597e46-5cc8-495d-95a1-f94235f2f9ed/kube-rbac-proxy/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.958511 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757d8h9wz_07597e46-5cc8-495d-95a1-f94235f2f9ed/manager/0.log" Oct 10 17:36:35 crc kubenswrapper[4788]: I1010 17:36:35.966104 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-749665c6b7-qx9vb_10351a76-d81b-4d40-935e-2dc3eb1b3960/kube-rbac-proxy/0.log" Oct 10 17:36:36 crc kubenswrapper[4788]: I1010 17:36:36.184912 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-65c468ddf8-db6bq_70d295ac-34a2-4d46-9d27-27c859268604/kube-rbac-proxy/0.log" Oct 10 17:36:36 crc kubenswrapper[4788]: I1010 17:36:36.367340 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-mxkqp_1a2d26de-5eca-4761-b78e-c259524b68f8/registry-server/0.log" Oct 10 17:36:36 crc kubenswrapper[4788]: I1010 17:36:36.440012 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-65c468ddf8-db6bq_70d295ac-34a2-4d46-9d27-27c859268604/operator/0.log" Oct 10 17:36:36 crc kubenswrapper[4788]: I1010 17:36:36.480221 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-kbw8c_d24a6fbc-3938-4612-a928-c0d97cba7fa8/kube-rbac-proxy/0.log" Oct 10 17:36:36 crc kubenswrapper[4788]: I1010 17:36:36.716890 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-4gq6b_d99a1777-7ea0-4072-b086-aabc93f57f71/kube-rbac-proxy/0.log" Oct 10 17:36:36 crc kubenswrapper[4788]: I1010 17:36:36.748489 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-kbw8c_d24a6fbc-3938-4612-a928-c0d97cba7fa8/manager/0.log" Oct 10 17:36:36 crc kubenswrapper[4788]: I1010 17:36:36.827433 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-4gq6b_d99a1777-7ea0-4072-b086-aabc93f57f71/manager/0.log" Oct 10 17:36:37 crc kubenswrapper[4788]: I1010 17:36:37.051230 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-b8vjv_7940c932-25d0-4f02-bcd5-e3441aa026c4/operator/0.log" Oct 10 17:36:37 crc kubenswrapper[4788]: I1010 17:36:37.059030 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-2z2xl_38f6b32d-a5f6-402f-8070-eb1252c7d669/kube-rbac-proxy/0.log" Oct 10 17:36:37 crc kubenswrapper[4788]: I1010 17:36:37.147635 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-2z2xl_38f6b32d-a5f6-402f-8070-eb1252c7d669/manager/0.log" Oct 10 17:36:38 crc kubenswrapper[4788]: I1010 17:36:38.001509 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-jvfjk_22f63d73-f623-4b6a-83d7-dff55231ecc7/kube-rbac-proxy/0.log" Oct 10 17:36:38 crc kubenswrapper[4788]: I1010 17:36:38.054357 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-7tlj8_ebfc2c03-b35c-41a7-9679-11677bca30b2/kube-rbac-proxy/0.log" Oct 10 17:36:38 crc kubenswrapper[4788]: I1010 17:36:38.184011 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-7tlj8_ebfc2c03-b35c-41a7-9679-11677bca30b2/manager/0.log" Oct 10 17:36:38 crc kubenswrapper[4788]: I1010 17:36:38.295302 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-hcq88_c2dbece7-c07a-4223-98a2-bb96f93b9e3c/kube-rbac-proxy/0.log" Oct 10 17:36:38 crc kubenswrapper[4788]: I1010 17:36:38.419848 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-jvfjk_22f63d73-f623-4b6a-83d7-dff55231ecc7/manager/0.log" Oct 10 17:36:38 crc kubenswrapper[4788]: I1010 17:36:38.483732 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-hcq88_c2dbece7-c07a-4223-98a2-bb96f93b9e3c/manager/0.log" Oct 10 17:36:38 crc kubenswrapper[4788]: I1010 17:36:38.671937 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-749665c6b7-qx9vb_10351a76-d81b-4d40-935e-2dc3eb1b3960/manager/0.log" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.228288 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gdl4x"] Oct 10 17:36:43 crc kubenswrapper[4788]: E1010 17:36:43.229574 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerName="extract-utilities" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.229620 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerName="extract-utilities" Oct 10 17:36:43 crc kubenswrapper[4788]: E1010 17:36:43.229653 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="registry-server" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.229661 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="registry-server" Oct 10 17:36:43 crc kubenswrapper[4788]: E1010 17:36:43.229686 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="extract-utilities" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.229695 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="extract-utilities" Oct 10 17:36:43 crc kubenswrapper[4788]: E1010 17:36:43.229709 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="extract-content" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.229715 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="extract-content" Oct 10 17:36:43 crc kubenswrapper[4788]: E1010 17:36:43.229728 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerName="extract-content" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.229734 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerName="extract-content" Oct 10 17:36:43 crc kubenswrapper[4788]: E1010 17:36:43.229758 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerName="registry-server" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.229763 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerName="registry-server" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.230093 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba1217ab-5440-46ca-8b91-b263a85361b9" containerName="registry-server" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.230118 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="70bfd12c-3169-4c20-ae69-41f5421c3071" containerName="registry-server" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.232509 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.241625 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gdl4x"] Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.338829 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-catalog-content\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.339447 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-utilities\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.339626 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-779zz\" (UniqueName: \"kubernetes.io/projected/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-kube-api-access-779zz\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.441950 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-779zz\" (UniqueName: \"kubernetes.io/projected/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-kube-api-access-779zz\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.442107 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-catalog-content\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.442196 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-utilities\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.442595 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-catalog-content\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.442853 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-utilities\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.473457 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-779zz\" (UniqueName: \"kubernetes.io/projected/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-kube-api-access-779zz\") pod \"community-operators-gdl4x\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:43 crc kubenswrapper[4788]: I1010 17:36:43.560660 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:44 crc kubenswrapper[4788]: I1010 17:36:44.176519 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gdl4x"] Oct 10 17:36:44 crc kubenswrapper[4788]: I1010 17:36:44.219836 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdl4x" event={"ID":"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698","Type":"ContainerStarted","Data":"94d804edfad8b90a80a3aae9a41d7430301ed81f781d03552dbc31aa0bf92a19"} Oct 10 17:36:45 crc kubenswrapper[4788]: I1010 17:36:45.233066 4788 generic.go:334] "Generic (PLEG): container finished" podID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerID="63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023" exitCode=0 Oct 10 17:36:45 crc kubenswrapper[4788]: I1010 17:36:45.233168 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdl4x" event={"ID":"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698","Type":"ContainerDied","Data":"63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023"} Oct 10 17:36:46 crc kubenswrapper[4788]: I1010 17:36:46.271251 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdl4x" event={"ID":"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698","Type":"ContainerStarted","Data":"349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d"} Oct 10 17:36:47 crc kubenswrapper[4788]: I1010 17:36:47.281424 4788 generic.go:334] "Generic (PLEG): container finished" podID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerID="349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d" exitCode=0 Oct 10 17:36:47 crc kubenswrapper[4788]: I1010 17:36:47.281503 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdl4x" event={"ID":"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698","Type":"ContainerDied","Data":"349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d"} Oct 10 17:36:48 crc kubenswrapper[4788]: I1010 17:36:48.297956 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdl4x" event={"ID":"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698","Type":"ContainerStarted","Data":"81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456"} Oct 10 17:36:48 crc kubenswrapper[4788]: I1010 17:36:48.341362 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gdl4x" podStartSLOduration=2.8587450690000002 podStartE2EDuration="5.341333327s" podCreationTimestamp="2025-10-10 17:36:43 +0000 UTC" firstStartedPulling="2025-10-10 17:36:45.237320896 +0000 UTC m=+10307.687036444" lastFinishedPulling="2025-10-10 17:36:47.719909164 +0000 UTC m=+10310.169624702" observedRunningTime="2025-10-10 17:36:48.3297519 +0000 UTC m=+10310.779467488" watchObservedRunningTime="2025-10-10 17:36:48.341333327 +0000 UTC m=+10310.791048875" Oct 10 17:36:53 crc kubenswrapper[4788]: I1010 17:36:53.560825 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:53 crc kubenswrapper[4788]: I1010 17:36:53.561316 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:53 crc kubenswrapper[4788]: I1010 17:36:53.628939 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:54 crc kubenswrapper[4788]: I1010 17:36:54.423396 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:54 crc kubenswrapper[4788]: I1010 17:36:54.471850 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gdl4x"] Oct 10 17:36:56 crc kubenswrapper[4788]: I1010 17:36:56.277657 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-cgbfx_6b2c9335-c98e-4808-b9cf-ecac89cd1a99/control-plane-machine-set-operator/0.log" Oct 10 17:36:56 crc kubenswrapper[4788]: I1010 17:36:56.393356 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gdl4x" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerName="registry-server" containerID="cri-o://81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456" gracePeriod=2 Oct 10 17:36:56 crc kubenswrapper[4788]: I1010 17:36:56.461855 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9bw5d_a46b6887-dafb-4cd1-a4f6-2a9602b84b89/kube-rbac-proxy/0.log" Oct 10 17:36:56 crc kubenswrapper[4788]: I1010 17:36:56.472858 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9bw5d_a46b6887-dafb-4cd1-a4f6-2a9602b84b89/machine-api-operator/0.log" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.002444 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.169519 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-utilities\") pod \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.169777 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-779zz\" (UniqueName: \"kubernetes.io/projected/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-kube-api-access-779zz\") pod \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.169834 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-catalog-content\") pod \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\" (UID: \"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698\") " Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.171413 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-utilities" (OuterVolumeSpecName: "utilities") pod "7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" (UID: "7bf3fb7b-8c3a-4fb9-b1b3-b04480675698"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.181399 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-kube-api-access-779zz" (OuterVolumeSpecName: "kube-api-access-779zz") pod "7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" (UID: "7bf3fb7b-8c3a-4fb9-b1b3-b04480675698"). InnerVolumeSpecName "kube-api-access-779zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.221711 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" (UID: "7bf3fb7b-8c3a-4fb9-b1b3-b04480675698"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.272939 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.272985 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-779zz\" (UniqueName: \"kubernetes.io/projected/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-kube-api-access-779zz\") on node \"crc\" DevicePath \"\"" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.272996 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.405587 4788 generic.go:334] "Generic (PLEG): container finished" podID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerID="81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456" exitCode=0 Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.405642 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdl4x" event={"ID":"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698","Type":"ContainerDied","Data":"81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456"} Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.405659 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdl4x" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.405676 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdl4x" event={"ID":"7bf3fb7b-8c3a-4fb9-b1b3-b04480675698","Type":"ContainerDied","Data":"94d804edfad8b90a80a3aae9a41d7430301ed81f781d03552dbc31aa0bf92a19"} Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.405698 4788 scope.go:117] "RemoveContainer" containerID="81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.432326 4788 scope.go:117] "RemoveContainer" containerID="349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.472230 4788 scope.go:117] "RemoveContainer" containerID="63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.478077 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gdl4x"] Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.487882 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gdl4x"] Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.520588 4788 scope.go:117] "RemoveContainer" containerID="81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456" Oct 10 17:36:57 crc kubenswrapper[4788]: E1010 17:36:57.521406 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456\": container with ID starting with 81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456 not found: ID does not exist" containerID="81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.521459 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456"} err="failed to get container status \"81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456\": rpc error: code = NotFound desc = could not find container \"81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456\": container with ID starting with 81b1fd3b3cab8d84cfa1cd99850729ea36cca7cba725e9b47db8d90096b1a456 not found: ID does not exist" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.521491 4788 scope.go:117] "RemoveContainer" containerID="349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d" Oct 10 17:36:57 crc kubenswrapper[4788]: E1010 17:36:57.521960 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d\": container with ID starting with 349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d not found: ID does not exist" containerID="349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.521985 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d"} err="failed to get container status \"349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d\": rpc error: code = NotFound desc = could not find container \"349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d\": container with ID starting with 349ddc3d4a55b70f150b26d7b76801345e98bb671b69d892bbd97a68f2798f8d not found: ID does not exist" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.521998 4788 scope.go:117] "RemoveContainer" containerID="63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023" Oct 10 17:36:57 crc kubenswrapper[4788]: E1010 17:36:57.522506 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023\": container with ID starting with 63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023 not found: ID does not exist" containerID="63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023" Oct 10 17:36:57 crc kubenswrapper[4788]: I1010 17:36:57.522565 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023"} err="failed to get container status \"63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023\": rpc error: code = NotFound desc = could not find container \"63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023\": container with ID starting with 63d6a8f32ffd5312df53a279f2190f3aebcfdd005e7128539f94671a3a235023 not found: ID does not exist" Oct 10 17:36:58 crc kubenswrapper[4788]: I1010 17:36:58.245243 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" path="/var/lib/kubelet/pods/7bf3fb7b-8c3a-4fb9-b1b3-b04480675698/volumes" Oct 10 17:37:09 crc kubenswrapper[4788]: I1010 17:37:09.717113 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-lhrlm_1ab62119-f153-45ab-8b1a-073493aa5f4b/cert-manager-controller/0.log" Oct 10 17:37:09 crc kubenswrapper[4788]: I1010 17:37:09.817121 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-mgv2n_8c24f8c3-1871-44fb-a84c-e4d5b5885604/cert-manager-cainjector/0.log" Oct 10 17:37:09 crc kubenswrapper[4788]: I1010 17:37:09.907385 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-5jvst_921d9dbb-d20b-4a32-b7ad-ff9ab5e694da/cert-manager-webhook/0.log" Oct 10 17:37:23 crc kubenswrapper[4788]: I1010 17:37:23.635183 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-d78sc_c5e5ab85-6606-4e44-bc7b-8fe2df836f49/nmstate-console-plugin/0.log" Oct 10 17:37:23 crc kubenswrapper[4788]: I1010 17:37:23.816522 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lg5x2_34af49c2-e879-487f-b92e-c178ea679f4d/nmstate-handler/0.log" Oct 10 17:37:23 crc kubenswrapper[4788]: I1010 17:37:23.922190 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-7crrv_c7329403-a7f9-45bb-93e9-7fdc91953d06/kube-rbac-proxy/0.log" Oct 10 17:37:23 crc kubenswrapper[4788]: I1010 17:37:23.924135 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-7crrv_c7329403-a7f9-45bb-93e9-7fdc91953d06/nmstate-metrics/0.log" Oct 10 17:37:24 crc kubenswrapper[4788]: I1010 17:37:24.125722 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-k6czn_a6752f9a-c8ce-4c2b-b234-b5229f430a5e/nmstate-operator/0.log" Oct 10 17:37:24 crc kubenswrapper[4788]: I1010 17:37:24.206844 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-5q4md_d3b269c8-9fb0-4982-a512-eb5e91d37a97/nmstate-webhook/0.log" Oct 10 17:37:40 crc kubenswrapper[4788]: I1010 17:37:40.733372 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-g8jk6_02700c1c-08fb-45a3-956b-bdbc6ac4e18a/kube-rbac-proxy/0.log" Oct 10 17:37:40 crc kubenswrapper[4788]: I1010 17:37:40.914735 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-frr-files/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.160260 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-g8jk6_02700c1c-08fb-45a3-956b-bdbc6ac4e18a/controller/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.176797 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-reloader/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.196294 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-frr-files/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.263902 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-metrics/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.358374 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-reloader/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.502657 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-reloader/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.537352 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-metrics/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.567850 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-metrics/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.579480 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-frr-files/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.723217 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-frr-files/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.748115 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-metrics/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.762795 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/cp-reloader/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.778555 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/controller/0.log" Oct 10 17:37:41 crc kubenswrapper[4788]: I1010 17:37:41.989395 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/kube-rbac-proxy/0.log" Oct 10 17:37:42 crc kubenswrapper[4788]: I1010 17:37:42.001422 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/frr-metrics/0.log" Oct 10 17:37:42 crc kubenswrapper[4788]: I1010 17:37:42.011895 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/kube-rbac-proxy-frr/0.log" Oct 10 17:37:42 crc kubenswrapper[4788]: I1010 17:37:42.204711 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/reloader/0.log" Oct 10 17:37:42 crc kubenswrapper[4788]: I1010 17:37:42.280744 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-5bj2r_85b09555-e996-4aee-8882-5e2bf28f4d1b/frr-k8s-webhook-server/0.log" Oct 10 17:37:42 crc kubenswrapper[4788]: I1010 17:37:42.535339 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b8b68658d-sn5zq_68109e8b-88f9-48ad-b162-282574174516/manager/0.log" Oct 10 17:37:42 crc kubenswrapper[4788]: I1010 17:37:42.717768 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d8576d6b4-hr28x_2f9b7183-67aa-4cea-83c6-59a32ff3f027/webhook-server/0.log" Oct 10 17:37:42 crc kubenswrapper[4788]: I1010 17:37:42.805971 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8gzh5_ef3cd7ba-2155-4457-a62c-63660545e0cc/kube-rbac-proxy/0.log" Oct 10 17:37:43 crc kubenswrapper[4788]: I1010 17:37:43.937876 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8gzh5_ef3cd7ba-2155-4457-a62c-63660545e0cc/speaker/0.log" Oct 10 17:37:45 crc kubenswrapper[4788]: I1010 17:37:45.631570 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v9bn2_22cedde9-131c-445d-832d-8063a6d3adf6/frr/0.log" Oct 10 17:37:57 crc kubenswrapper[4788]: I1010 17:37:57.405452 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl_c6f44e2f-cadb-4a7c-baa1-2305298df8c5/util/0.log" Oct 10 17:37:57 crc kubenswrapper[4788]: I1010 17:37:57.599609 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl_c6f44e2f-cadb-4a7c-baa1-2305298df8c5/util/0.log" Oct 10 17:37:57 crc kubenswrapper[4788]: I1010 17:37:57.667676 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl_c6f44e2f-cadb-4a7c-baa1-2305298df8c5/pull/0.log" Oct 10 17:37:57 crc kubenswrapper[4788]: I1010 17:37:57.694182 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl_c6f44e2f-cadb-4a7c-baa1-2305298df8c5/pull/0.log" Oct 10 17:37:57 crc kubenswrapper[4788]: I1010 17:37:57.875914 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl_c6f44e2f-cadb-4a7c-baa1-2305298df8c5/util/0.log" Oct 10 17:37:57 crc kubenswrapper[4788]: I1010 17:37:57.935165 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl_c6f44e2f-cadb-4a7c-baa1-2305298df8c5/pull/0.log" Oct 10 17:37:57 crc kubenswrapper[4788]: I1010 17:37:57.936472 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69pqsjl_c6f44e2f-cadb-4a7c-baa1-2305298df8c5/extract/0.log" Oct 10 17:37:58 crc kubenswrapper[4788]: I1010 17:37:58.087815 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4_60d1e35d-ed52-4de4-8aef-b10d13e2d5c9/util/0.log" Oct 10 17:37:58 crc kubenswrapper[4788]: I1010 17:37:58.281562 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4_60d1e35d-ed52-4de4-8aef-b10d13e2d5c9/pull/0.log" Oct 10 17:37:58 crc kubenswrapper[4788]: I1010 17:37:58.345087 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4_60d1e35d-ed52-4de4-8aef-b10d13e2d5c9/util/0.log" Oct 10 17:37:58 crc kubenswrapper[4788]: I1010 17:37:58.351536 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4_60d1e35d-ed52-4de4-8aef-b10d13e2d5c9/pull/0.log" Oct 10 17:37:58 crc kubenswrapper[4788]: I1010 17:37:58.558663 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4_60d1e35d-ed52-4de4-8aef-b10d13e2d5c9/util/0.log" Oct 10 17:37:58 crc kubenswrapper[4788]: I1010 17:37:58.571976 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4_60d1e35d-ed52-4de4-8aef-b10d13e2d5c9/extract/0.log" Oct 10 17:37:58 crc kubenswrapper[4788]: I1010 17:37:58.618080 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xgsb4_60d1e35d-ed52-4de4-8aef-b10d13e2d5c9/pull/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.063166 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8_65ef9ce9-fa0d-4d4d-b31b-5e1392189319/util/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.198739 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8_65ef9ce9-fa0d-4d4d-b31b-5e1392189319/pull/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.198818 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8_65ef9ce9-fa0d-4d4d-b31b-5e1392189319/pull/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.249470 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8_65ef9ce9-fa0d-4d4d-b31b-5e1392189319/util/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.406521 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.406637 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.467550 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8_65ef9ce9-fa0d-4d4d-b31b-5e1392189319/pull/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.477941 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8_65ef9ce9-fa0d-4d4d-b31b-5e1392189319/util/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.514963 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d9c5q8_65ef9ce9-fa0d-4d4d-b31b-5e1392189319/extract/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.696362 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84nzb_471efb95-ace7-444f-8786-f9eed7209eba/extract-utilities/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.908865 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84nzb_471efb95-ace7-444f-8786-f9eed7209eba/extract-content/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.921529 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84nzb_471efb95-ace7-444f-8786-f9eed7209eba/extract-content/0.log" Oct 10 17:37:59 crc kubenswrapper[4788]: I1010 17:37:59.928849 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84nzb_471efb95-ace7-444f-8786-f9eed7209eba/extract-utilities/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.134776 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84nzb_471efb95-ace7-444f-8786-f9eed7209eba/extract-utilities/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.191906 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84nzb_471efb95-ace7-444f-8786-f9eed7209eba/extract-content/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.360948 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zd7kn_0edf0140-7a81-4ce5-bbb2-3cfebd76db36/extract-utilities/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.536346 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-84nzb_471efb95-ace7-444f-8786-f9eed7209eba/registry-server/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.604101 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zd7kn_0edf0140-7a81-4ce5-bbb2-3cfebd76db36/extract-utilities/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.609812 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zd7kn_0edf0140-7a81-4ce5-bbb2-3cfebd76db36/extract-content/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.654355 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zd7kn_0edf0140-7a81-4ce5-bbb2-3cfebd76db36/extract-content/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.862042 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zd7kn_0edf0140-7a81-4ce5-bbb2-3cfebd76db36/extract-utilities/0.log" Oct 10 17:38:00 crc kubenswrapper[4788]: I1010 17:38:00.965106 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zd7kn_0edf0140-7a81-4ce5-bbb2-3cfebd76db36/extract-content/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.017522 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd_33653f49-1164-45f1-90b9-d22e4c531b5e/util/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.189995 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-zd7kn_0edf0140-7a81-4ce5-bbb2-3cfebd76db36/registry-server/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.302013 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd_33653f49-1164-45f1-90b9-d22e4c531b5e/util/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.305473 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd_33653f49-1164-45f1-90b9-d22e4c531b5e/pull/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.314801 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd_33653f49-1164-45f1-90b9-d22e4c531b5e/pull/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.504454 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd_33653f49-1164-45f1-90b9-d22e4c531b5e/util/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.509520 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd_33653f49-1164-45f1-90b9-d22e4c531b5e/pull/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.513376 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9htdd_33653f49-1164-45f1-90b9-d22e4c531b5e/extract/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.553957 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-pd5nn_04d1b9a1-3e55-480c-aab6-257b2095fd7d/marketplace-operator/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.694843 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rskph_85d1a4d9-5c27-447a-922e-73c4edac912c/extract-utilities/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.926921 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rskph_85d1a4d9-5c27-447a-922e-73c4edac912c/extract-utilities/0.log" Oct 10 17:38:01 crc kubenswrapper[4788]: I1010 17:38:01.954046 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rskph_85d1a4d9-5c27-447a-922e-73c4edac912c/extract-content/0.log" Oct 10 17:38:02 crc kubenswrapper[4788]: I1010 17:38:02.009262 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rskph_85d1a4d9-5c27-447a-922e-73c4edac912c/extract-content/0.log" Oct 10 17:38:02 crc kubenswrapper[4788]: I1010 17:38:02.174359 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rskph_85d1a4d9-5c27-447a-922e-73c4edac912c/extract-utilities/0.log" Oct 10 17:38:02 crc kubenswrapper[4788]: I1010 17:38:02.199622 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rskph_85d1a4d9-5c27-447a-922e-73c4edac912c/extract-content/0.log" Oct 10 17:38:02 crc kubenswrapper[4788]: I1010 17:38:02.207759 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9598_dcf1e71b-3d94-4450-ae39-ea10deac5dde/extract-utilities/0.log" Oct 10 17:38:02 crc kubenswrapper[4788]: I1010 17:38:02.444125 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9598_dcf1e71b-3d94-4450-ae39-ea10deac5dde/extract-utilities/0.log" Oct 10 17:38:02 crc kubenswrapper[4788]: I1010 17:38:02.446361 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9598_dcf1e71b-3d94-4450-ae39-ea10deac5dde/extract-content/0.log" Oct 10 17:38:02 crc kubenswrapper[4788]: I1010 17:38:02.483788 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9598_dcf1e71b-3d94-4450-ae39-ea10deac5dde/extract-content/0.log" Oct 10 17:38:02 crc kubenswrapper[4788]: I1010 17:38:02.580615 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rskph_85d1a4d9-5c27-447a-922e-73c4edac912c/registry-server/0.log" Oct 10 17:38:03 crc kubenswrapper[4788]: I1010 17:38:03.066066 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9598_dcf1e71b-3d94-4450-ae39-ea10deac5dde/extract-utilities/0.log" Oct 10 17:38:03 crc kubenswrapper[4788]: I1010 17:38:03.097181 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9598_dcf1e71b-3d94-4450-ae39-ea10deac5dde/extract-content/0.log" Oct 10 17:38:04 crc kubenswrapper[4788]: I1010 17:38:04.832447 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9598_dcf1e71b-3d94-4450-ae39-ea10deac5dde/registry-server/0.log" Oct 10 17:38:16 crc kubenswrapper[4788]: I1010 17:38:16.504617 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-78srq_deb5ec8e-ffbc-4cb8-8404-3e0903aa4934/prometheus-operator/0.log" Oct 10 17:38:17 crc kubenswrapper[4788]: I1010 17:38:17.801490 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5d4477dcb7-rjwnz_640304e5-e29f-4190-8615-b5e1f71d4f31/prometheus-operator-admission-webhook/0.log" Oct 10 17:38:17 crc kubenswrapper[4788]: I1010 17:38:17.835850 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5d4477dcb7-v52g8_3bf651f1-a6e1-458b-87f7-b5faec152462/prometheus-operator-admission-webhook/0.log" Oct 10 17:38:17 crc kubenswrapper[4788]: I1010 17:38:17.984365 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-qjr28_982e6e19-775c-45bf-86ff-47233ac227e2/operator/0.log" Oct 10 17:38:18 crc kubenswrapper[4788]: I1010 17:38:18.062128 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-p5kcr_83e92919-6e65-42d4-9a5d-ce7c78831457/perses-operator/0.log" Oct 10 17:38:24 crc kubenswrapper[4788]: E1010 17:38:24.328995 4788 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.129:41278->38.102.83.129:41727: write tcp 38.102.83.129:41278->38.102.83.129:41727: write: broken pipe Oct 10 17:38:29 crc kubenswrapper[4788]: I1010 17:38:29.412390 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:38:29 crc kubenswrapper[4788]: I1010 17:38:29.413447 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:38:30 crc kubenswrapper[4788]: E1010 17:38:30.684976 4788 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.129:41652->38.102.83.129:41727: write tcp 38.102.83.129:41652->38.102.83.129:41727: write: broken pipe Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.405982 4788 patch_prober.go:28] interesting pod/machine-config-daemon-dxfdf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.406916 4788 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.406993 4788 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.408475 4788 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac"} pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.408567 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" containerName="machine-config-daemon" containerID="cri-o://ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" gracePeriod=600 Oct 10 17:38:59 crc kubenswrapper[4788]: E1010 17:38:59.538166 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.819311 4788 generic.go:334] "Generic (PLEG): container finished" podID="2419d2de-214a-4a13-b941-2acd571f0360" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" exitCode=0 Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.819415 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" event={"ID":"2419d2de-214a-4a13-b941-2acd571f0360","Type":"ContainerDied","Data":"ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac"} Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.819498 4788 scope.go:117] "RemoveContainer" containerID="93046d31ffbebfcb70b32194820289c869f35af289dc70cd1e47ca5fb7e4b2bf" Oct 10 17:38:59 crc kubenswrapper[4788]: I1010 17:38:59.820990 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:38:59 crc kubenswrapper[4788]: E1010 17:38:59.821621 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:39:13 crc kubenswrapper[4788]: I1010 17:39:13.233332 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:39:13 crc kubenswrapper[4788]: E1010 17:39:13.234055 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:39:23 crc kubenswrapper[4788]: I1010 17:39:23.015390 4788 scope.go:117] "RemoveContainer" containerID="b2b06f6839d441a349f3551457279ed2f62ed62442cb06f2b62336e513948757" Oct 10 17:39:27 crc kubenswrapper[4788]: I1010 17:39:27.234089 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:39:27 crc kubenswrapper[4788]: E1010 17:39:27.235245 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:39:41 crc kubenswrapper[4788]: I1010 17:39:41.233982 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:39:41 crc kubenswrapper[4788]: E1010 17:39:41.236055 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:39:53 crc kubenswrapper[4788]: I1010 17:39:53.233899 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:39:53 crc kubenswrapper[4788]: E1010 17:39:53.236401 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:40:05 crc kubenswrapper[4788]: I1010 17:40:05.235339 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:40:05 crc kubenswrapper[4788]: E1010 17:40:05.236936 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:40:17 crc kubenswrapper[4788]: I1010 17:40:17.234526 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:40:17 crc kubenswrapper[4788]: E1010 17:40:17.235387 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.478519 4788 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9m4nt"] Oct 10 17:40:24 crc kubenswrapper[4788]: E1010 17:40:24.479459 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerName="extract-utilities" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.479473 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerName="extract-utilities" Oct 10 17:40:24 crc kubenswrapper[4788]: E1010 17:40:24.479492 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerName="extract-content" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.479498 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerName="extract-content" Oct 10 17:40:24 crc kubenswrapper[4788]: E1010 17:40:24.479520 4788 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerName="registry-server" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.479526 4788 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerName="registry-server" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.479733 4788 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bf3fb7b-8c3a-4fb9-b1b3-b04480675698" containerName="registry-server" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.481382 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.503117 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9m4nt"] Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.676778 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-catalog-content\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.678273 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-688nb\" (UniqueName: \"kubernetes.io/projected/186e989d-0486-47a6-bc58-2170b0683488-kube-api-access-688nb\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.678561 4788 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-utilities\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.781103 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-utilities\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.781264 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-catalog-content\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.781388 4788 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-688nb\" (UniqueName: \"kubernetes.io/projected/186e989d-0486-47a6-bc58-2170b0683488-kube-api-access-688nb\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.782109 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-utilities\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.782321 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-catalog-content\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.814068 4788 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-688nb\" (UniqueName: \"kubernetes.io/projected/186e989d-0486-47a6-bc58-2170b0683488-kube-api-access-688nb\") pod \"certified-operators-9m4nt\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:24 crc kubenswrapper[4788]: I1010 17:40:24.823642 4788 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:25 crc kubenswrapper[4788]: I1010 17:40:25.516309 4788 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9m4nt"] Oct 10 17:40:25 crc kubenswrapper[4788]: I1010 17:40:25.826796 4788 generic.go:334] "Generic (PLEG): container finished" podID="186e989d-0486-47a6-bc58-2170b0683488" containerID="f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534" exitCode=0 Oct 10 17:40:25 crc kubenswrapper[4788]: I1010 17:40:25.827184 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4nt" event={"ID":"186e989d-0486-47a6-bc58-2170b0683488","Type":"ContainerDied","Data":"f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534"} Oct 10 17:40:25 crc kubenswrapper[4788]: I1010 17:40:25.827215 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4nt" event={"ID":"186e989d-0486-47a6-bc58-2170b0683488","Type":"ContainerStarted","Data":"1d760b11602b57afb2d97842c5e7910c77426e48f364c48642b42bdcde113a10"} Oct 10 17:40:25 crc kubenswrapper[4788]: I1010 17:40:25.833547 4788 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 10 17:40:27 crc kubenswrapper[4788]: I1010 17:40:27.869517 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4nt" event={"ID":"186e989d-0486-47a6-bc58-2170b0683488","Type":"ContainerStarted","Data":"0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621"} Oct 10 17:40:28 crc kubenswrapper[4788]: I1010 17:40:28.884252 4788 generic.go:334] "Generic (PLEG): container finished" podID="186e989d-0486-47a6-bc58-2170b0683488" containerID="0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621" exitCode=0 Oct 10 17:40:28 crc kubenswrapper[4788]: I1010 17:40:28.884373 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4nt" event={"ID":"186e989d-0486-47a6-bc58-2170b0683488","Type":"ContainerDied","Data":"0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621"} Oct 10 17:40:29 crc kubenswrapper[4788]: I1010 17:40:29.899629 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4nt" event={"ID":"186e989d-0486-47a6-bc58-2170b0683488","Type":"ContainerStarted","Data":"3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2"} Oct 10 17:40:29 crc kubenswrapper[4788]: I1010 17:40:29.934109 4788 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9m4nt" podStartSLOduration=2.38461459 podStartE2EDuration="5.934074467s" podCreationTimestamp="2025-10-10 17:40:24 +0000 UTC" firstStartedPulling="2025-10-10 17:40:25.833330573 +0000 UTC m=+10528.283046121" lastFinishedPulling="2025-10-10 17:40:29.38279044 +0000 UTC m=+10531.832505998" observedRunningTime="2025-10-10 17:40:29.924197886 +0000 UTC m=+10532.373913494" watchObservedRunningTime="2025-10-10 17:40:29.934074467 +0000 UTC m=+10532.383790055" Oct 10 17:40:32 crc kubenswrapper[4788]: I1010 17:40:32.234516 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:40:32 crc kubenswrapper[4788]: E1010 17:40:32.235204 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:40:34 crc kubenswrapper[4788]: I1010 17:40:34.824914 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:34 crc kubenswrapper[4788]: I1010 17:40:34.825535 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:34 crc kubenswrapper[4788]: I1010 17:40:34.890903 4788 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:35 crc kubenswrapper[4788]: I1010 17:40:35.035094 4788 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:35 crc kubenswrapper[4788]: I1010 17:40:35.129859 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9m4nt"] Oct 10 17:40:36 crc kubenswrapper[4788]: I1010 17:40:36.988882 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9m4nt" podUID="186e989d-0486-47a6-bc58-2170b0683488" containerName="registry-server" containerID="cri-o://3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2" gracePeriod=2 Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.540984 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.592649 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-688nb\" (UniqueName: \"kubernetes.io/projected/186e989d-0486-47a6-bc58-2170b0683488-kube-api-access-688nb\") pod \"186e989d-0486-47a6-bc58-2170b0683488\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.592949 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-utilities\") pod \"186e989d-0486-47a6-bc58-2170b0683488\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.593168 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-catalog-content\") pod \"186e989d-0486-47a6-bc58-2170b0683488\" (UID: \"186e989d-0486-47a6-bc58-2170b0683488\") " Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.594025 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-utilities" (OuterVolumeSpecName: "utilities") pod "186e989d-0486-47a6-bc58-2170b0683488" (UID: "186e989d-0486-47a6-bc58-2170b0683488"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.611087 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/186e989d-0486-47a6-bc58-2170b0683488-kube-api-access-688nb" (OuterVolumeSpecName: "kube-api-access-688nb") pod "186e989d-0486-47a6-bc58-2170b0683488" (UID: "186e989d-0486-47a6-bc58-2170b0683488"). InnerVolumeSpecName "kube-api-access-688nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.663519 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "186e989d-0486-47a6-bc58-2170b0683488" (UID: "186e989d-0486-47a6-bc58-2170b0683488"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.695325 4788 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.695370 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-688nb\" (UniqueName: \"kubernetes.io/projected/186e989d-0486-47a6-bc58-2170b0683488-kube-api-access-688nb\") on node \"crc\" DevicePath \"\"" Oct 10 17:40:37 crc kubenswrapper[4788]: I1010 17:40:37.695385 4788 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/186e989d-0486-47a6-bc58-2170b0683488-utilities\") on node \"crc\" DevicePath \"\"" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.003524 4788 generic.go:334] "Generic (PLEG): container finished" podID="186e989d-0486-47a6-bc58-2170b0683488" containerID="3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2" exitCode=0 Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.003597 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4nt" event={"ID":"186e989d-0486-47a6-bc58-2170b0683488","Type":"ContainerDied","Data":"3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2"} Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.003678 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9m4nt" event={"ID":"186e989d-0486-47a6-bc58-2170b0683488","Type":"ContainerDied","Data":"1d760b11602b57afb2d97842c5e7910c77426e48f364c48642b42bdcde113a10"} Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.003701 4788 scope.go:117] "RemoveContainer" containerID="3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.006414 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9m4nt" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.034061 4788 scope.go:117] "RemoveContainer" containerID="0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.061879 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9m4nt"] Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.066549 4788 scope.go:117] "RemoveContainer" containerID="f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.076550 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9m4nt"] Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.120251 4788 scope.go:117] "RemoveContainer" containerID="3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2" Oct 10 17:40:38 crc kubenswrapper[4788]: E1010 17:40:38.121156 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2\": container with ID starting with 3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2 not found: ID does not exist" containerID="3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.121189 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2"} err="failed to get container status \"3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2\": rpc error: code = NotFound desc = could not find container \"3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2\": container with ID starting with 3747e4c9b1f0c5126eb57992b05ec4329a970c52ed71ebcaa1fea63739bd0fa2 not found: ID does not exist" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.121214 4788 scope.go:117] "RemoveContainer" containerID="0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621" Oct 10 17:40:38 crc kubenswrapper[4788]: E1010 17:40:38.121717 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621\": container with ID starting with 0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621 not found: ID does not exist" containerID="0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.121749 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621"} err="failed to get container status \"0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621\": rpc error: code = NotFound desc = could not find container \"0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621\": container with ID starting with 0797a94f9f963c248c128bc86d287f5b89a238c9b1749eef0e0ee30c814fe621 not found: ID does not exist" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.121763 4788 scope.go:117] "RemoveContainer" containerID="f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534" Oct 10 17:40:38 crc kubenswrapper[4788]: E1010 17:40:38.122359 4788 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534\": container with ID starting with f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534 not found: ID does not exist" containerID="f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.122384 4788 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534"} err="failed to get container status \"f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534\": rpc error: code = NotFound desc = could not find container \"f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534\": container with ID starting with f2fabe6a833cfe4982d86c92be6085827370f85799fd52daf2297ef17dfdd534 not found: ID does not exist" Oct 10 17:40:38 crc kubenswrapper[4788]: I1010 17:40:38.246620 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="186e989d-0486-47a6-bc58-2170b0683488" path="/var/lib/kubelet/pods/186e989d-0486-47a6-bc58-2170b0683488/volumes" Oct 10 17:40:39 crc kubenswrapper[4788]: I1010 17:40:39.017275 4788 generic.go:334] "Generic (PLEG): container finished" podID="5c789567-43c2-4f52-b62d-cd6b956a4f47" containerID="1c3f4cf0fd24e46294ad698717488da6c2e6d637e8fe2db32e108f640e723e52" exitCode=0 Oct 10 17:40:39 crc kubenswrapper[4788]: I1010 17:40:39.017371 4788 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7xjww/must-gather-wfqvz" event={"ID":"5c789567-43c2-4f52-b62d-cd6b956a4f47","Type":"ContainerDied","Data":"1c3f4cf0fd24e46294ad698717488da6c2e6d637e8fe2db32e108f640e723e52"} Oct 10 17:40:39 crc kubenswrapper[4788]: I1010 17:40:39.020091 4788 scope.go:117] "RemoveContainer" containerID="1c3f4cf0fd24e46294ad698717488da6c2e6d637e8fe2db32e108f640e723e52" Oct 10 17:40:39 crc kubenswrapper[4788]: I1010 17:40:39.998526 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7xjww_must-gather-wfqvz_5c789567-43c2-4f52-b62d-cd6b956a4f47/gather/0.log" Oct 10 17:40:46 crc kubenswrapper[4788]: I1010 17:40:46.234935 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:40:46 crc kubenswrapper[4788]: E1010 17:40:46.235894 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:40:48 crc kubenswrapper[4788]: I1010 17:40:48.879018 4788 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7xjww/must-gather-wfqvz"] Oct 10 17:40:48 crc kubenswrapper[4788]: I1010 17:40:48.880162 4788 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7xjww/must-gather-wfqvz" podUID="5c789567-43c2-4f52-b62d-cd6b956a4f47" containerName="copy" containerID="cri-o://5f56aa7cdfe98aed70e4952da11818dcc32258966b2a3f9dbefd4e2dcad912fb" gracePeriod=2 Oct 10 17:40:48 crc kubenswrapper[4788]: I1010 17:40:48.900714 4788 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7xjww/must-gather-wfqvz"] Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.187399 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7xjww_must-gather-wfqvz_5c789567-43c2-4f52-b62d-cd6b956a4f47/copy/0.log" Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.188158 4788 generic.go:334] "Generic (PLEG): container finished" podID="5c789567-43c2-4f52-b62d-cd6b956a4f47" containerID="5f56aa7cdfe98aed70e4952da11818dcc32258966b2a3f9dbefd4e2dcad912fb" exitCode=143 Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.386235 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7xjww_must-gather-wfqvz_5c789567-43c2-4f52-b62d-cd6b956a4f47/copy/0.log" Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.386897 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.530852 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4nwf\" (UniqueName: \"kubernetes.io/projected/5c789567-43c2-4f52-b62d-cd6b956a4f47-kube-api-access-l4nwf\") pod \"5c789567-43c2-4f52-b62d-cd6b956a4f47\" (UID: \"5c789567-43c2-4f52-b62d-cd6b956a4f47\") " Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.530982 4788 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5c789567-43c2-4f52-b62d-cd6b956a4f47-must-gather-output\") pod \"5c789567-43c2-4f52-b62d-cd6b956a4f47\" (UID: \"5c789567-43c2-4f52-b62d-cd6b956a4f47\") " Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.541489 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c789567-43c2-4f52-b62d-cd6b956a4f47-kube-api-access-l4nwf" (OuterVolumeSpecName: "kube-api-access-l4nwf") pod "5c789567-43c2-4f52-b62d-cd6b956a4f47" (UID: "5c789567-43c2-4f52-b62d-cd6b956a4f47"). InnerVolumeSpecName "kube-api-access-l4nwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.633695 4788 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4nwf\" (UniqueName: \"kubernetes.io/projected/5c789567-43c2-4f52-b62d-cd6b956a4f47-kube-api-access-l4nwf\") on node \"crc\" DevicePath \"\"" Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.781425 4788 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c789567-43c2-4f52-b62d-cd6b956a4f47-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5c789567-43c2-4f52-b62d-cd6b956a4f47" (UID: "5c789567-43c2-4f52-b62d-cd6b956a4f47"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 10 17:40:49 crc kubenswrapper[4788]: I1010 17:40:49.843234 4788 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5c789567-43c2-4f52-b62d-cd6b956a4f47-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 10 17:40:50 crc kubenswrapper[4788]: I1010 17:40:50.203991 4788 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7xjww_must-gather-wfqvz_5c789567-43c2-4f52-b62d-cd6b956a4f47/copy/0.log" Oct 10 17:40:50 crc kubenswrapper[4788]: I1010 17:40:50.207232 4788 scope.go:117] "RemoveContainer" containerID="5f56aa7cdfe98aed70e4952da11818dcc32258966b2a3f9dbefd4e2dcad912fb" Oct 10 17:40:50 crc kubenswrapper[4788]: I1010 17:40:50.207702 4788 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7xjww/must-gather-wfqvz" Oct 10 17:40:50 crc kubenswrapper[4788]: I1010 17:40:50.265632 4788 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c789567-43c2-4f52-b62d-cd6b956a4f47" path="/var/lib/kubelet/pods/5c789567-43c2-4f52-b62d-cd6b956a4f47/volumes" Oct 10 17:40:50 crc kubenswrapper[4788]: I1010 17:40:50.269808 4788 scope.go:117] "RemoveContainer" containerID="1c3f4cf0fd24e46294ad698717488da6c2e6d637e8fe2db32e108f640e723e52" Oct 10 17:40:59 crc kubenswrapper[4788]: I1010 17:40:59.234773 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:40:59 crc kubenswrapper[4788]: E1010 17:40:59.235665 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:41:11 crc kubenswrapper[4788]: I1010 17:41:11.235641 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:41:11 crc kubenswrapper[4788]: E1010 17:41:11.236479 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:41:24 crc kubenswrapper[4788]: I1010 17:41:24.234468 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:41:24 crc kubenswrapper[4788]: E1010 17:41:24.235307 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:41:36 crc kubenswrapper[4788]: I1010 17:41:36.234543 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:41:36 crc kubenswrapper[4788]: E1010 17:41:36.235324 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:41:48 crc kubenswrapper[4788]: I1010 17:41:48.243891 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:41:48 crc kubenswrapper[4788]: E1010 17:41:48.244818 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:42:03 crc kubenswrapper[4788]: I1010 17:42:03.234239 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:42:03 crc kubenswrapper[4788]: E1010 17:42:03.234922 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:42:16 crc kubenswrapper[4788]: I1010 17:42:16.233775 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:42:16 crc kubenswrapper[4788]: E1010 17:42:16.235521 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" Oct 10 17:42:30 crc kubenswrapper[4788]: I1010 17:42:30.235055 4788 scope.go:117] "RemoveContainer" containerID="ae08933860e75b02745ceacc3393d0a79829084b0c492050bd1cbb066796a4ac" Oct 10 17:42:30 crc kubenswrapper[4788]: E1010 17:42:30.236047 4788 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dxfdf_openshift-machine-config-operator(2419d2de-214a-4a13-b941-2acd571f0360)\"" pod="openshift-machine-config-operator/machine-config-daemon-dxfdf" podUID="2419d2de-214a-4a13-b941-2acd571f0360" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515072242633024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015072242633017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015072215362016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015072215363015461 5ustar corecore